[ 487.448958] env[63021]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=63021) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 487.449465] env[63021]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=63021) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 487.449465] env[63021]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=63021) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 487.449691] env[63021]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 487.546817] env[63021]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=63021) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 487.556912] env[63021]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=63021) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 488.157529] env[63021]: INFO nova.virt.driver [None req-326eb2ea-8483-4cd6-a372-00f6b404d1b9 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 488.227485] env[63021]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 488.227686] env[63021]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 488.227751] env[63021]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=63021) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 491.379391] env[63021]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-5eca1f8f-e940-4503-8e07-095daf0085de {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 491.394937] env[63021]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=63021) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 491.395132] env[63021]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-37336484-b5da-409f-84f6-7017897f23f2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 491.426290] env[63021]: INFO oslo_vmware.api [-] Successfully established new session; session ID is f62b9. [ 491.426556] env[63021]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.199s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 491.427018] env[63021]: INFO nova.virt.vmwareapi.driver [None req-326eb2ea-8483-4cd6-a372-00f6b404d1b9 None None] VMware vCenter version: 7.0.3 [ 491.430415] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56231543-e876-4027-b5e5-7e25e40ea3e5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 491.447591] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9440ca9-5589-455c-b02a-e28d90c6aac8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 491.453270] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16e38d80-15ac-4f95-9955-2e4ba31d4405 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 491.459791] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99a382c5-b094-47fb-b3db-170cf216acaa {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 491.472611] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dde26f8b-ef03-4a61-91ad-63dcc8a0f102 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 491.478427] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-955c207e-e18b-49a2-b2ac-bfa50f215622 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 491.508089] env[63021]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-16c7407c-d4ff-49af-bc43-8e4126fc37e0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 491.513235] env[63021]: DEBUG nova.virt.vmwareapi.driver [None req-326eb2ea-8483-4cd6-a372-00f6b404d1b9 None None] Extension org.openstack.compute already exists. {{(pid=63021) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:225}} [ 491.515909] env[63021]: INFO nova.compute.provider_config [None req-326eb2ea-8483-4cd6-a372-00f6b404d1b9 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 492.019171] env[63021]: DEBUG nova.context [None req-326eb2ea-8483-4cd6-a372-00f6b404d1b9 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),d53e36ca-1c25-45ef-ad41-06b382c56e65(cell1) {{(pid=63021) load_cells /opt/stack/nova/nova/context.py:464}} [ 492.021559] env[63021]: DEBUG oslo_concurrency.lockutils [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 492.021800] env[63021]: DEBUG oslo_concurrency.lockutils [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 492.022574] env[63021]: DEBUG oslo_concurrency.lockutils [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 492.023061] env[63021]: DEBUG oslo_concurrency.lockutils [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] Acquiring lock "d53e36ca-1c25-45ef-ad41-06b382c56e65" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 492.023287] env[63021]: DEBUG oslo_concurrency.lockutils [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] Lock "d53e36ca-1c25-45ef-ad41-06b382c56e65" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 492.024458] env[63021]: DEBUG oslo_concurrency.lockutils [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] Lock "d53e36ca-1c25-45ef-ad41-06b382c56e65" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 492.044687] env[63021]: INFO dbcounter [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] Registered counter for database nova_cell0 [ 492.052803] env[63021]: INFO dbcounter [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] Registered counter for database nova_cell1 [ 492.056165] env[63021]: DEBUG oslo_db.sqlalchemy.engines [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=63021) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 492.056533] env[63021]: DEBUG oslo_db.sqlalchemy.engines [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=63021) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 492.061496] env[63021]: ERROR nova.db.main.api [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 492.061496] env[63021]: result = function(*args, **kwargs) [ 492.061496] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 492.061496] env[63021]: return func(*args, **kwargs) [ 492.061496] env[63021]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 492.061496] env[63021]: result = fn(*args, **kwargs) [ 492.061496] env[63021]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 492.061496] env[63021]: return f(*args, **kwargs) [ 492.061496] env[63021]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 492.061496] env[63021]: return db.service_get_minimum_version(context, binaries) [ 492.061496] env[63021]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 492.061496] env[63021]: _check_db_access() [ 492.061496] env[63021]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 492.061496] env[63021]: stacktrace = ''.join(traceback.format_stack()) [ 492.061496] env[63021]: [ 492.062319] env[63021]: ERROR nova.db.main.api [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 492.062319] env[63021]: result = function(*args, **kwargs) [ 492.062319] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 492.062319] env[63021]: return func(*args, **kwargs) [ 492.062319] env[63021]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 492.062319] env[63021]: result = fn(*args, **kwargs) [ 492.062319] env[63021]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 492.062319] env[63021]: return f(*args, **kwargs) [ 492.062319] env[63021]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 492.062319] env[63021]: return db.service_get_minimum_version(context, binaries) [ 492.062319] env[63021]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 492.062319] env[63021]: _check_db_access() [ 492.062319] env[63021]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 492.062319] env[63021]: stacktrace = ''.join(traceback.format_stack()) [ 492.062319] env[63021]: [ 492.062756] env[63021]: WARNING nova.objects.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 492.062846] env[63021]: WARNING nova.objects.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] Failed to get minimum service version for cell d53e36ca-1c25-45ef-ad41-06b382c56e65 [ 492.063273] env[63021]: DEBUG oslo_concurrency.lockutils [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] Acquiring lock "singleton_lock" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 492.063435] env[63021]: DEBUG oslo_concurrency.lockutils [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] Acquired lock "singleton_lock" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 492.063677] env[63021]: DEBUG oslo_concurrency.lockutils [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] Releasing lock "singleton_lock" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 492.064007] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] Full set of CONF: {{(pid=63021) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 492.064159] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ******************************************************************************** {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 492.064288] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] Configuration options gathered from: {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 492.064427] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 492.064617] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 492.064748] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ================================================================================ {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 492.064950] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] allow_resize_to_same_host = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.065135] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] arq_binding_timeout = 300 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.065316] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] backdoor_port = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.065395] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] backdoor_socket = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.065584] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] block_device_allocate_retries = 60 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.065750] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] block_device_allocate_retries_interval = 3 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.065917] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cert = self.pem {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.066096] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.066267] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] compute_monitors = [] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.066433] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] config_dir = [] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.066592] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] config_drive_format = iso9660 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.066729] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.066892] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] config_source = [] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.067068] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] console_host = devstack {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.067236] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] control_exchange = nova {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.067391] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cpu_allocation_ratio = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.067548] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] daemon = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.067713] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] debug = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.067866] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] default_access_ip_network_name = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.068040] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] default_availability_zone = nova {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.068196] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] default_ephemeral_format = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.068355] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] default_green_pool_size = 1000 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.068623] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.068793] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] default_schedule_zone = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.068950] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] disk_allocation_ratio = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.069129] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] enable_new_services = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.069310] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] enabled_apis = ['osapi_compute'] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.069474] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] enabled_ssl_apis = [] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.069633] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] flat_injected = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.069790] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] force_config_drive = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.069946] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] force_raw_images = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.070159] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] graceful_shutdown_timeout = 5 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.070331] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] heal_instance_info_cache_interval = 60 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.070546] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] host = cpu-1 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.070721] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.070884] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] initial_disk_allocation_ratio = 1.0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.071063] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] initial_ram_allocation_ratio = 1.0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.071277] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.071440] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] instance_build_timeout = 0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.071631] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] instance_delete_interval = 300 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.071800] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] instance_format = [instance: %(uuid)s] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.071963] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] instance_name_template = instance-%08x {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.072137] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] instance_usage_audit = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.072305] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] instance_usage_audit_period = month {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.072475] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.072633] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] instances_path = /opt/stack/data/nova/instances {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.072796] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] internal_service_availability_zone = internal {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.072953] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] key = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.073123] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] live_migration_retry_count = 30 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.073293] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] log_color = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.073457] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] log_config_append = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.073617] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.073772] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] log_dir = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.073926] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] log_file = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.074063] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] log_options = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.074260] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] log_rotate_interval = 1 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.074390] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] log_rotate_interval_type = days {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.074574] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] log_rotation_type = none {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.074711] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.074838] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.075010] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.075182] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.075313] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.075478] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] long_rpc_timeout = 1800 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.075633] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] max_concurrent_builds = 10 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.075791] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] max_concurrent_live_migrations = 1 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.075945] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] max_concurrent_snapshots = 5 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.076112] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] max_local_block_devices = 3 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.076270] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] max_logfile_count = 30 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.076424] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] max_logfile_size_mb = 200 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.076580] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] maximum_instance_delete_attempts = 5 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.076745] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] metadata_listen = 0.0.0.0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.076907] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] metadata_listen_port = 8775 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.077083] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] metadata_workers = 2 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.077250] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] migrate_max_retries = -1 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.077412] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] mkisofs_cmd = genisoimage {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.077637] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] my_block_storage_ip = 10.180.1.21 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.077775] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] my_ip = 10.180.1.21 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.077937] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] network_allocate_retries = 0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.078158] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.078334] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] osapi_compute_listen = 0.0.0.0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.078497] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] osapi_compute_listen_port = 8774 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.078659] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] osapi_compute_unique_server_name_scope = {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.078819] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] osapi_compute_workers = 2 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.078979] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] password_length = 12 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.079144] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] periodic_enable = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.079350] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] periodic_fuzzy_delay = 60 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.079463] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] pointer_model = usbtablet {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.079622] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] preallocate_images = none {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.079775] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] publish_errors = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.079900] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] pybasedir = /opt/stack/nova {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.080091] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ram_allocation_ratio = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.080252] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] rate_limit_burst = 0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.080424] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] rate_limit_except_level = CRITICAL {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.080580] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] rate_limit_interval = 0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.080734] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] reboot_timeout = 0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.080887] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] reclaim_instance_interval = 0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.081048] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] record = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.081215] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] reimage_timeout_per_gb = 60 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.081376] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] report_interval = 120 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.081530] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] rescue_timeout = 0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.081681] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] reserved_host_cpus = 0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.081830] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] reserved_host_disk_mb = 0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.081979] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] reserved_host_memory_mb = 512 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.082142] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] reserved_huge_pages = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.082297] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] resize_confirm_window = 0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.082457] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] resize_fs_using_block_device = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.082605] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] resume_guests_state_on_host_boot = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.082765] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.082917] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] rpc_response_timeout = 60 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.083082] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] run_external_periodic_tasks = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.083249] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] running_deleted_instance_action = reap {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.083409] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] running_deleted_instance_poll_interval = 1800 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.083557] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] running_deleted_instance_timeout = 0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.083705] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] scheduler_instance_sync_interval = 120 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.083862] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] service_down_time = 720 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.084031] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] servicegroup_driver = db {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.084185] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] shell_completion = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.084338] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] shelved_offload_time = 0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.084488] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] shelved_poll_interval = 3600 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.084646] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] shutdown_timeout = 0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.084799] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] source_is_ipv6 = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.084953] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ssl_only = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.085298] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.085548] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] sync_power_state_interval = 600 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.085714] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] sync_power_state_pool_size = 1000 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.085915] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] syslog_log_facility = LOG_USER {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.086092] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] tempdir = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.086275] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] timeout_nbd = 10 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.086441] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] transport_url = **** {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.086598] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] update_resources_interval = 0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.086755] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] use_cow_images = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.086908] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] use_eventlog = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.087073] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] use_journal = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.087232] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] use_json = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.087384] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] use_rootwrap_daemon = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.087536] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] use_stderr = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.088098] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] use_syslog = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.088098] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vcpu_pin_set = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.088098] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vif_plugging_is_fatal = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.088184] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vif_plugging_timeout = 300 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.088344] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] virt_mkfs = [] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.088504] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] volume_usage_poll_interval = 0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.088660] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] watch_log_file = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.088823] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] web = /usr/share/spice-html5 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 492.089012] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.089181] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.089353] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.089525] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_concurrency.disable_process_locking = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.090133] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.090335] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.090513] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.090688] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.090861] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.091039] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.091229] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api.auth_strategy = keystone {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.091396] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api.compute_link_prefix = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.091936] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.091936] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api.dhcp_domain = novalocal {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.091936] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api.enable_instance_password = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.092085] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api.glance_link_prefix = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.092226] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.092411] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.092557] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api.instance_list_per_project_cells = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.092720] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api.list_records_by_skipping_down_cells = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.092877] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api.local_metadata_per_cell = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.093054] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api.max_limit = 1000 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.093220] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api.metadata_cache_expiration = 15 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.093392] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api.neutron_default_tenant_id = default {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.093555] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api.response_validation = warn {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.093719] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api.use_neutron_default_nets = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.093881] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.094098] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.094296] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.094470] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.094639] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api.vendordata_dynamic_targets = [] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.094801] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api.vendordata_jsonfile_path = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.094973] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.095180] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.backend = dogpile.cache.memcached {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.095350] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.backend_argument = **** {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.095519] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.config_prefix = cache.oslo {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.095685] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.dead_timeout = 60.0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.095847] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.debug_cache_backend = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.096024] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.enable_retry_client = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.096175] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.enable_socket_keepalive = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.096343] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.enabled = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.096505] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.enforce_fips_mode = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.096667] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.expiration_time = 600 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.096827] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.hashclient_retry_attempts = 2 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.096990] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.hashclient_retry_delay = 1.0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.097174] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.memcache_dead_retry = 300 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.097337] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.memcache_password = **** {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.097498] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.097660] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.097819] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.memcache_pool_maxsize = 10 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.097993] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.098187] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.memcache_sasl_enabled = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.098373] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.098543] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.memcache_socket_timeout = 1.0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.098700] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.memcache_username = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.098864] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.proxies = [] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.099054] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.redis_db = 0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.099225] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.redis_password = **** {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.099399] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.redis_sentinel_service_name = mymaster {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.099571] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.099743] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.redis_server = localhost:6379 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.099907] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.redis_socket_timeout = 1.0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.100106] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.redis_username = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.100270] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.retry_attempts = 2 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.100436] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.retry_delay = 0.0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.100597] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.socket_keepalive_count = 1 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.100755] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.socket_keepalive_idle = 1 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.100915] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.socket_keepalive_interval = 1 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.101091] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.tls_allowed_ciphers = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.101254] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.tls_cafile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.101404] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.tls_certfile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.101561] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.tls_enabled = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.101715] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cache.tls_keyfile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.101883] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cinder.auth_section = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.102067] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cinder.auth_type = password {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.102235] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cinder.cafile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.102416] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cinder.catalog_info = volumev3::publicURL {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.102573] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cinder.certfile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.102734] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cinder.collect_timing = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.102893] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cinder.cross_az_attach = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.103063] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cinder.debug = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.103226] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cinder.endpoint_template = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.103389] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cinder.http_retries = 3 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.103551] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cinder.insecure = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.103706] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cinder.keyfile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.103873] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cinder.os_region_name = RegionOne {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.104044] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cinder.split_loggers = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.104210] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cinder.timeout = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.104381] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.104539] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] compute.cpu_dedicated_set = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.104695] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] compute.cpu_shared_set = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.104858] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] compute.image_type_exclude_list = [] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.105026] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.105193] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] compute.max_concurrent_disk_ops = 0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.105357] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] compute.max_disk_devices_to_attach = -1 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.105517] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.105725] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.105845] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] compute.resource_provider_association_refresh = 300 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.105997] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.106170] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] compute.shutdown_retry_interval = 10 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.106349] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.106525] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] conductor.workers = 2 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.106703] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] console.allowed_origins = [] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.106865] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] console.ssl_ciphers = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.107046] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] console.ssl_minimum_version = default {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.107218] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] consoleauth.enforce_session_timeout = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.107388] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] consoleauth.token_ttl = 600 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.107557] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cyborg.cafile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.107716] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cyborg.certfile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.107878] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cyborg.collect_timing = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.108071] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cyborg.connect_retries = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.108245] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cyborg.connect_retry_delay = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.108406] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cyborg.endpoint_override = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.108570] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cyborg.insecure = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.108730] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cyborg.keyfile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.108888] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cyborg.max_version = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.109054] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cyborg.min_version = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.109215] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cyborg.region_name = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.109377] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cyborg.retriable_status_codes = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.109534] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cyborg.service_name = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.109701] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cyborg.service_type = accelerator {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.109861] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cyborg.split_loggers = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.110028] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cyborg.status_code_retries = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.110212] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cyborg.status_code_retry_delay = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.110375] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cyborg.timeout = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.110559] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.110722] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] cyborg.version = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.110899] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] database.backend = sqlalchemy {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.111080] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] database.connection = **** {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.111258] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] database.connection_debug = 0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.111420] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] database.connection_parameters = {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.111583] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] database.connection_recycle_time = 3600 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.111746] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] database.connection_trace = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.111905] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] database.db_inc_retry_interval = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.112077] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] database.db_max_retries = 20 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.112242] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] database.db_max_retry_interval = 10 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.112401] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] database.db_retry_interval = 1 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.112561] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] database.max_overflow = 50 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.112720] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] database.max_pool_size = 5 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.112878] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] database.max_retries = 10 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.113059] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.113223] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] database.mysql_wsrep_sync_wait = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.113381] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] database.pool_timeout = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.113540] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] database.retry_interval = 10 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.113696] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] database.slave_connection = **** {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.113854] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] database.sqlite_synchronous = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.114021] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] database.use_db_reconnect = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.114202] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api_database.backend = sqlalchemy {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.114371] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api_database.connection = **** {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.114533] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api_database.connection_debug = 0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.114699] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api_database.connection_parameters = {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.114860] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api_database.connection_recycle_time = 3600 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.115030] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api_database.connection_trace = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.115196] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api_database.db_inc_retry_interval = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.115360] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api_database.db_max_retries = 20 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.115521] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api_database.db_max_retry_interval = 10 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.115683] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api_database.db_retry_interval = 1 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.115839] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api_database.max_overflow = 50 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.116040] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api_database.max_pool_size = 5 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.116166] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api_database.max_retries = 10 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.116333] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.116493] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.116648] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api_database.pool_timeout = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.116806] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api_database.retry_interval = 10 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.116966] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api_database.slave_connection = **** {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.117137] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] api_database.sqlite_synchronous = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.117315] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] devices.enabled_mdev_types = [] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.117489] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.117662] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ephemeral_storage_encryption.default_format = luks {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.117826] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ephemeral_storage_encryption.enabled = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.118013] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.118207] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] glance.api_servers = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.118376] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] glance.cafile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.118540] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] glance.certfile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.118702] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] glance.collect_timing = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.118860] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] glance.connect_retries = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.119029] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] glance.connect_retry_delay = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.119195] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] glance.debug = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.119372] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] glance.default_trusted_certificate_ids = [] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.119526] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] glance.enable_certificate_validation = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.119686] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] glance.enable_rbd_download = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.119842] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] glance.endpoint_override = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.120013] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] glance.insecure = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.120220] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] glance.keyfile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.120384] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] glance.max_version = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.120578] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] glance.min_version = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.120749] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] glance.num_retries = 3 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.120923] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] glance.rbd_ceph_conf = {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.121101] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] glance.rbd_connect_timeout = 5 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.121278] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] glance.rbd_pool = {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.121445] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] glance.rbd_user = {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.121605] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] glance.region_name = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.121768] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] glance.retriable_status_codes = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.121928] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] glance.service_name = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.122106] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] glance.service_type = image {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.122272] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] glance.split_loggers = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.122439] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] glance.status_code_retries = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.122596] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] glance.status_code_retry_delay = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.122754] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] glance.timeout = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.122931] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.123105] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] glance.verify_glance_signatures = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.123272] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] glance.version = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.123441] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] guestfs.debug = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.123610] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] mks.enabled = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.123966] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.124199] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] image_cache.manager_interval = 2400 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.124382] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] image_cache.precache_concurrency = 1 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.124555] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] image_cache.remove_unused_base_images = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.124726] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.124894] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.125082] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] image_cache.subdirectory_name = _base {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.125266] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ironic.api_max_retries = 60 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.125432] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ironic.api_retry_interval = 2 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.125595] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ironic.auth_section = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.125760] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ironic.auth_type = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.125919] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ironic.cafile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.126092] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ironic.certfile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.126258] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ironic.collect_timing = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.126423] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ironic.conductor_group = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.126582] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ironic.connect_retries = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.126740] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ironic.connect_retry_delay = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.126898] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ironic.endpoint_override = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.127070] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ironic.insecure = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.127239] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ironic.keyfile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.127398] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ironic.max_version = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.127555] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ironic.min_version = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.127722] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ironic.peer_list = [] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.127881] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ironic.region_name = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.128075] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ironic.retriable_status_codes = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.128268] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ironic.serial_console_state_timeout = 10 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.128433] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ironic.service_name = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.128604] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ironic.service_type = baremetal {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.128766] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ironic.shard = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.128929] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ironic.split_loggers = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.129101] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ironic.status_code_retries = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.129265] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ironic.status_code_retry_delay = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.129422] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ironic.timeout = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.129604] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.129764] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ironic.version = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.129947] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.130166] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] key_manager.fixed_key = **** {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.130374] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.130533] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] barbican.barbican_api_version = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.130691] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] barbican.barbican_endpoint = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.130862] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] barbican.barbican_endpoint_type = public {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.131031] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] barbican.barbican_region_name = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.131221] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] barbican.cafile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.131405] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] barbican.certfile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.131572] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] barbican.collect_timing = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.131733] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] barbican.insecure = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.131891] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] barbican.keyfile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.132084] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] barbican.number_of_retries = 60 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.132273] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] barbican.retry_delay = 1 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.132442] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] barbican.send_service_user_token = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.132604] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] barbican.split_loggers = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.132765] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] barbican.timeout = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.132923] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] barbican.verify_ssl = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.133093] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] barbican.verify_ssl_path = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.133263] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] barbican_service_user.auth_section = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.133425] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] barbican_service_user.auth_type = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.133580] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] barbican_service_user.cafile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.133736] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] barbican_service_user.certfile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.133895] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] barbican_service_user.collect_timing = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.134065] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] barbican_service_user.insecure = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.134227] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] barbican_service_user.keyfile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.134389] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] barbican_service_user.split_loggers = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.134544] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] barbican_service_user.timeout = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.134707] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vault.approle_role_id = **** {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.134864] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vault.approle_secret_id = **** {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.135046] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vault.kv_mountpoint = secret {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.135212] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vault.kv_path = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.135375] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vault.kv_version = 2 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.135536] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vault.namespace = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.135692] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vault.root_token_id = **** {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.135847] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vault.ssl_ca_crt_file = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.136025] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vault.timeout = 60.0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.136223] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vault.use_ssl = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.136396] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.136566] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] keystone.auth_section = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.136732] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] keystone.auth_type = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.136924] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] keystone.cafile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.137103] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] keystone.certfile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.137270] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] keystone.collect_timing = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.137428] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] keystone.connect_retries = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.137586] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] keystone.connect_retry_delay = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.137743] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] keystone.endpoint_override = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.137901] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] keystone.insecure = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.138069] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] keystone.keyfile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.138227] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] keystone.max_version = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.138381] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] keystone.min_version = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.138536] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] keystone.region_name = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.138692] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] keystone.retriable_status_codes = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.138849] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] keystone.service_name = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.139023] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] keystone.service_type = identity {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.139189] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] keystone.split_loggers = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.139345] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] keystone.status_code_retries = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.139500] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] keystone.status_code_retry_delay = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.139655] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] keystone.timeout = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.139832] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.139991] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] keystone.version = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.140265] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.connection_uri = {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.140433] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.cpu_mode = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.140598] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.cpu_model_extra_flags = [] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.140768] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.cpu_models = [] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.140941] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.cpu_power_governor_high = performance {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.141125] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.cpu_power_governor_low = powersave {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.141322] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.cpu_power_management = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.141501] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.141666] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.device_detach_attempts = 8 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.141829] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.device_detach_timeout = 20 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.141994] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.disk_cachemodes = [] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.142179] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.disk_prefix = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.142378] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.enabled_perf_events = [] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.142523] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.file_backed_memory = 0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.142688] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.gid_maps = [] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.142845] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.hw_disk_discard = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.142999] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.hw_machine_type = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.143184] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.images_rbd_ceph_conf = {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.143352] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.143551] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.143727] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.images_rbd_glance_store_name = {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.143896] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.images_rbd_pool = rbd {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.144095] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.images_type = default {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.144276] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.images_volume_group = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.144437] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.inject_key = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.144596] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.inject_partition = -2 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.144755] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.inject_password = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.144915] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.iscsi_iface = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.145087] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.iser_use_multipath = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.145256] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.live_migration_bandwidth = 0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.145420] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.145583] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.live_migration_downtime = 500 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.145743] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.145900] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.146070] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.live_migration_inbound_addr = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.146268] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.146410] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.live_migration_permit_post_copy = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.146553] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.live_migration_scheme = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.146726] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.live_migration_timeout_action = abort {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.146887] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.live_migration_tunnelled = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.147057] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.live_migration_uri = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.147220] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.live_migration_with_native_tls = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.147380] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.max_queues = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.147545] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.147788] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.147957] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.nfs_mount_options = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.148332] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.148518] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.148685] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.num_iser_scan_tries = 5 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.148848] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.num_memory_encrypted_guests = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.149025] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.149195] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.num_pcie_ports = 0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.149373] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.num_volume_scan_tries = 5 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.149539] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.pmem_namespaces = [] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.149700] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.quobyte_client_cfg = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.149984] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.150196] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.rbd_connect_timeout = 5 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.150372] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.150538] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.150698] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.rbd_secret_uuid = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.150857] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.rbd_user = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.151030] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.151222] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.remote_filesystem_transport = ssh {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.151398] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.rescue_image_id = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.151593] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.rescue_kernel_id = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.151758] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.rescue_ramdisk_id = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.151927] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.152115] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.rx_queue_size = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.152300] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.smbfs_mount_options = {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.152598] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.152786] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.snapshot_compression = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.152950] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.snapshot_image_format = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.153187] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.153383] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.sparse_logical_volumes = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.153550] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.swtpm_enabled = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.153719] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.swtpm_group = tss {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.153884] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.swtpm_user = tss {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.154061] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.sysinfo_serial = unique {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.154225] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.tb_cache_size = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.154379] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.tx_queue_size = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.154542] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.uid_maps = [] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.154700] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.use_virtio_for_bridges = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.154866] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.virt_type = kvm {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.155042] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.volume_clear = zero {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.155206] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.volume_clear_size = 0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.155369] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.volume_use_multipath = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.155520] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.vzstorage_cache_path = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.155683] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.155845] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.vzstorage_mount_group = qemu {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.156014] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.vzstorage_mount_opts = [] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.156226] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.156503] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.156676] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.vzstorage_mount_user = stack {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.156838] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.157023] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] neutron.auth_section = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.157195] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] neutron.auth_type = password {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.157356] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] neutron.cafile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.157511] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] neutron.certfile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.157667] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] neutron.collect_timing = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.157821] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] neutron.connect_retries = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.157976] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] neutron.connect_retry_delay = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.158155] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] neutron.default_floating_pool = public {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.158313] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] neutron.endpoint_override = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.158472] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] neutron.extension_sync_interval = 600 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.158629] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] neutron.http_retries = 3 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.158786] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] neutron.insecure = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.158937] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] neutron.keyfile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.159102] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] neutron.max_version = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.159270] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.159424] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] neutron.min_version = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.159590] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] neutron.ovs_bridge = br-int {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.159753] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] neutron.physnets = [] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.159920] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] neutron.region_name = RegionOne {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.160130] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] neutron.retriable_status_codes = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.160333] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] neutron.service_metadata_proxy = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.160538] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] neutron.service_name = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.160788] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] neutron.service_type = network {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.160964] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] neutron.split_loggers = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.161200] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] neutron.status_code_retries = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.161389] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] neutron.status_code_retry_delay = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.161565] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] neutron.timeout = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.161745] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.161904] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] neutron.version = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.162087] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] notifications.bdms_in_notifications = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.162268] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] notifications.default_level = INFO {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.162453] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] notifications.notification_format = unversioned {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.162618] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] notifications.notify_on_state_change = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.162796] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.162973] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] pci.alias = [] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.163158] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] pci.device_spec = [] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.163325] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] pci.report_in_placement = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.163495] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.auth_section = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.163666] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.auth_type = password {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.163832] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.163987] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.cafile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.164187] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.certfile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.164358] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.collect_timing = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.164516] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.connect_retries = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.164672] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.connect_retry_delay = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.164826] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.default_domain_id = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.164982] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.default_domain_name = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.165154] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.domain_id = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.165310] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.domain_name = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.165464] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.endpoint_override = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.165620] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.insecure = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.165774] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.keyfile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.165928] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.max_version = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.166092] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.min_version = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.166262] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.password = **** {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.166425] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.project_domain_id = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.166601] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.project_domain_name = Default {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.166746] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.project_id = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.166914] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.project_name = service {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.167097] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.region_name = RegionOne {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.167264] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.retriable_status_codes = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.167420] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.service_name = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.167584] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.service_type = placement {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.167742] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.split_loggers = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.167898] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.status_code_retries = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.168134] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.status_code_retry_delay = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.168292] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.system_scope = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.168455] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.timeout = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.168611] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.trust_id = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.168765] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.user_domain_id = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.168931] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.user_domain_name = Default {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.169101] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.user_id = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.169280] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.username = nova {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.169456] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.169615] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] placement.version = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.169789] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] quota.cores = 20 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.169950] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] quota.count_usage_from_placement = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.170156] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.170344] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] quota.injected_file_content_bytes = 10240 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.170514] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] quota.injected_file_path_length = 255 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.170677] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] quota.injected_files = 5 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.170847] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] quota.instances = 10 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.171009] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] quota.key_pairs = 100 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.171181] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] quota.metadata_items = 128 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.171347] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] quota.ram = 51200 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.171509] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] quota.recheck_quota = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.171675] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] quota.server_group_members = 10 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.171839] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] quota.server_groups = 10 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.172026] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.172223] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.172393] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] scheduler.image_metadata_prefilter = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.172555] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.172715] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] scheduler.max_attempts = 3 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.172878] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] scheduler.max_placement_results = 1000 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.173052] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.173218] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] scheduler.query_placement_for_image_type_support = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.173381] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.173551] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] scheduler.workers = 2 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.173723] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.173893] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.174083] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.174257] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.174422] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.174587] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.174749] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.174938] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.175118] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] filter_scheduler.host_subset_size = 1 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.175287] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.175451] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.175616] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.175779] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] filter_scheduler.isolated_hosts = [] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.175942] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] filter_scheduler.isolated_images = [] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.176140] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.176339] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.176486] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.176693] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] filter_scheduler.pci_in_placement = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.176811] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.176973] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.177150] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.177316] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.177478] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.177636] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.177796] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] filter_scheduler.track_instance_changes = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.177969] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.178176] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] metrics.required = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.178354] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] metrics.weight_multiplier = 1.0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.178516] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.178680] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] metrics.weight_setting = [] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.178993] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.179182] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] serial_console.enabled = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.179362] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] serial_console.port_range = 10000:20000 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.179531] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.179696] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.179899] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] serial_console.serialproxy_port = 6083 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.180122] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] service_user.auth_section = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.180316] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] service_user.auth_type = password {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.180481] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] service_user.cafile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.180637] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] service_user.certfile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.180794] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] service_user.collect_timing = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.180948] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] service_user.insecure = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.181119] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] service_user.keyfile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.181289] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] service_user.send_service_user_token = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.181448] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] service_user.split_loggers = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.181615] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] service_user.timeout = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.181781] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] spice.agent_enabled = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.181939] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] spice.enabled = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.182271] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.182469] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.182635] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] spice.html5proxy_port = 6082 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.182793] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] spice.image_compression = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.182949] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] spice.jpeg_compression = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.183117] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] spice.playback_compression = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.183282] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] spice.require_secure = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.183446] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] spice.server_listen = 127.0.0.1 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.183613] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.183769] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] spice.streaming_mode = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.183930] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] spice.zlib_compression = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.184129] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] upgrade_levels.baseapi = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.184319] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] upgrade_levels.compute = auto {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.184479] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] upgrade_levels.conductor = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.184633] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] upgrade_levels.scheduler = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.184796] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vendordata_dynamic_auth.auth_section = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.184954] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vendordata_dynamic_auth.auth_type = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.185122] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vendordata_dynamic_auth.cafile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.185284] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vendordata_dynamic_auth.certfile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.185442] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.185598] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vendordata_dynamic_auth.insecure = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.185753] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vendordata_dynamic_auth.keyfile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.185907] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.186071] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vendordata_dynamic_auth.timeout = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.186248] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vmware.api_retry_count = 10 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.186435] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vmware.ca_file = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.186574] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vmware.cache_prefix = devstack-image-cache {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.186740] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vmware.cluster_name = testcl1 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.186896] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vmware.connection_pool_size = 10 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.187059] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vmware.console_delay_seconds = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.187228] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vmware.datastore_regex = ^datastore.* {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.187430] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.187598] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vmware.host_password = **** {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.187759] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vmware.host_port = 443 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.187924] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vmware.host_username = administrator@vsphere.local {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.188140] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vmware.insecure = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.188328] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vmware.integration_bridge = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.188498] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vmware.maximum_objects = 100 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.188658] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vmware.pbm_default_policy = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.188816] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vmware.pbm_enabled = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.188971] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vmware.pbm_wsdl_location = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.189152] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.189312] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vmware.serial_port_proxy_uri = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.189466] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vmware.serial_port_service_uri = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.189629] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vmware.task_poll_interval = 0.5 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.189794] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vmware.use_linked_clone = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.189957] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vmware.vnc_keymap = en-us {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.190161] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vmware.vnc_port = 5900 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.190341] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vmware.vnc_port_total = 10000 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.190529] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vnc.auth_schemes = ['none'] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.190705] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vnc.enabled = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.190997] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.191196] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.191368] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vnc.novncproxy_port = 6080 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.191544] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vnc.server_listen = 127.0.0.1 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.191712] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.191870] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vnc.vencrypt_ca_certs = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.192049] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vnc.vencrypt_client_cert = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.192229] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vnc.vencrypt_client_key = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.192411] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.192573] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] workarounds.disable_deep_image_inspection = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.192735] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.192896] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.193070] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.193238] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] workarounds.disable_rootwrap = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.193398] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] workarounds.enable_numa_live_migration = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.193558] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.193718] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.193878] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.194049] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] workarounds.libvirt_disable_apic = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.194212] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.194377] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.194537] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.194694] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.194854] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.195022] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.195187] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.195348] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.195508] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.195671] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.195853] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.196036] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] wsgi.client_socket_timeout = 900 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.196238] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] wsgi.default_pool_size = 1000 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.196412] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] wsgi.keep_alive = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.196980] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] wsgi.max_header_line = 16384 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.196980] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] wsgi.secure_proxy_ssl_header = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.196980] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] wsgi.ssl_ca_file = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.197108] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] wsgi.ssl_cert_file = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.197255] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] wsgi.ssl_key_file = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.197378] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] wsgi.tcp_keepidle = 600 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.197553] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.197719] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] zvm.ca_file = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.197878] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] zvm.cloud_connector_url = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.198219] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.198413] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] zvm.reachable_timeout = 300 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.198596] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_policy.enforce_new_defaults = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.198986] env[63021]: WARNING oslo_config.cfg [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 492.199185] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_policy.enforce_scope = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.199365] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_policy.policy_default_rule = default {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.199544] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.199715] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_policy.policy_file = policy.yaml {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.199892] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.200097] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.200297] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.200466] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.200631] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.200797] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.200974] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.201167] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] profiler.connection_string = messaging:// {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.201369] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] profiler.enabled = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.201525] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] profiler.es_doc_type = notification {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.201686] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] profiler.es_scroll_size = 10000 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.201852] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] profiler.es_scroll_time = 2m {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.202027] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] profiler.filter_error_trace = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.202192] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] profiler.hmac_keys = **** {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.202360] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] profiler.sentinel_service_name = mymaster {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.202524] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] profiler.socket_timeout = 0.1 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.202682] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] profiler.trace_requests = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.202840] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] profiler.trace_sqlalchemy = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.203405] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] profiler_jaeger.process_tags = {} {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.203405] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] profiler_jaeger.service_name_prefix = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.203405] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] profiler_otlp.service_name_prefix = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.203535] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] remote_debug.host = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.203633] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] remote_debug.port = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.203807] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.203968] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.204175] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.204352] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.204512] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.204672] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.204831] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.204991] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.205167] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.205339] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.205497] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.205665] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.205832] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.205999] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.206184] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.206352] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.206510] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.206683] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.206844] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.207011] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.207180] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.207351] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.207506] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.207665] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.207821] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.207978] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.208244] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.208472] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.208685] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.208891] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.ssl = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.209113] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.209329] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.209534] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.209745] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.209964] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.ssl_version = {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.210204] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.210449] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.210659] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_notifications.retry = -1 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.210882] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.211121] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_messaging_notifications.transport_url = **** {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.211381] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_limit.auth_section = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.211533] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_limit.auth_type = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.211732] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_limit.cafile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.211929] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_limit.certfile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.212149] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_limit.collect_timing = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.212357] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_limit.connect_retries = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.212554] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_limit.connect_retry_delay = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.212750] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_limit.endpoint_id = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.212943] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_limit.endpoint_override = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.213187] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_limit.insecure = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.213389] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_limit.keyfile = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.213584] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_limit.max_version = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.213774] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_limit.min_version = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.213972] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_limit.region_name = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.214186] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_limit.retriable_status_codes = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.214384] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_limit.service_name = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.214576] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_limit.service_type = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.214776] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_limit.split_loggers = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.214967] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_limit.status_code_retries = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.215176] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_limit.status_code_retry_delay = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.215371] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_limit.timeout = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.215571] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_limit.valid_interfaces = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.215757] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_limit.version = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.215957] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_reports.file_event_handler = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.216175] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.216376] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] oslo_reports.log_dir = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.216583] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.216777] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.216970] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.217207] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.217452] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.217611] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.217818] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.218034] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vif_plug_ovs_privileged.group = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.218286] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.218537] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.218773] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.219017] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] vif_plug_ovs_privileged.user = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.219280] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] os_vif_linux_bridge.flat_interface = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.219562] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.219810] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.220083] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.220287] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.220468] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.220641] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.220816] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.221022] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.221202] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] os_vif_ovs.isolate_vif = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.221376] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.221548] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.221717] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.221888] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] os_vif_ovs.ovsdb_interface = native {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.222063] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] os_vif_ovs.per_port_bridge = False {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.222239] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] privsep_osbrick.capabilities = [21] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.222403] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] privsep_osbrick.group = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.222859] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] privsep_osbrick.helper_command = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.222859] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.222859] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.223016] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] privsep_osbrick.user = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.223195] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.223356] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] nova_sys_admin.group = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.223512] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] nova_sys_admin.helper_command = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.223672] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.223833] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.223987] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] nova_sys_admin.user = None {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 492.224132] env[63021]: DEBUG oslo_service.service [None req-02763d3c-28ff-4298-aafe-08edcb72cca5 None None] ******************************************************************************** {{(pid=63021) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 492.224680] env[63021]: INFO nova.service [-] Starting compute node (version 30.1.0) [ 492.727863] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-1d0411b1-5d94-4c30-8c63-f930ce6d26d8 None None] Getting list of instances from cluster (obj){ [ 492.727863] env[63021]: value = "domain-c8" [ 492.727863] env[63021]: _type = "ClusterComputeResource" [ 492.727863] env[63021]: } {{(pid=63021) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 492.729287] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b9b854b-1c52-4d39-bdeb-1a584e92404c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.738063] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-1d0411b1-5d94-4c30-8c63-f930ce6d26d8 None None] Got total of 0 instances {{(pid=63021) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 492.738584] env[63021]: WARNING nova.virt.vmwareapi.driver [None req-1d0411b1-5d94-4c30-8c63-f930ce6d26d8 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 492.739042] env[63021]: INFO nova.virt.node [None req-1d0411b1-5d94-4c30-8c63-f930ce6d26d8 None None] Generated node identity 6ef6c041-2b0d-4bc3-8662-5d679a77a86d [ 492.739266] env[63021]: INFO nova.virt.node [None req-1d0411b1-5d94-4c30-8c63-f930ce6d26d8 None None] Wrote node identity 6ef6c041-2b0d-4bc3-8662-5d679a77a86d to /opt/stack/data/n-cpu-1/compute_id [ 493.242403] env[63021]: WARNING nova.compute.manager [None req-1d0411b1-5d94-4c30-8c63-f930ce6d26d8 None None] Compute nodes ['6ef6c041-2b0d-4bc3-8662-5d679a77a86d'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 494.249369] env[63021]: INFO nova.compute.manager [None req-1d0411b1-5d94-4c30-8c63-f930ce6d26d8 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 495.254729] env[63021]: WARNING nova.compute.manager [None req-1d0411b1-5d94-4c30-8c63-f930ce6d26d8 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 495.255095] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1d0411b1-5d94-4c30-8c63-f930ce6d26d8 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 495.255231] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1d0411b1-5d94-4c30-8c63-f930ce6d26d8 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 495.255348] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1d0411b1-5d94-4c30-8c63-f930ce6d26d8 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 495.255585] env[63021]: DEBUG nova.compute.resource_tracker [None req-1d0411b1-5d94-4c30-8c63-f930ce6d26d8 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63021) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 495.256428] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5590aa42-a59b-4b52-9f44-1f4af8c78f3a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 495.264844] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2abbcaaf-d918-418a-84c9-21526f5222b5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 495.278083] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e429f32c-2e74-4b71-bfcb-2cac61cef678 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 495.284294] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b10d0d3a-3086-49bf-be24-3d0c67ab3376 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 495.313468] env[63021]: DEBUG nova.compute.resource_tracker [None req-1d0411b1-5d94-4c30-8c63-f930ce6d26d8 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181521MB free_disk=139GB free_vcpus=48 pci_devices=None {{(pid=63021) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 495.313648] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1d0411b1-5d94-4c30-8c63-f930ce6d26d8 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 495.313815] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1d0411b1-5d94-4c30-8c63-f930ce6d26d8 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 495.815979] env[63021]: WARNING nova.compute.resource_tracker [None req-1d0411b1-5d94-4c30-8c63-f930ce6d26d8 None None] No compute node record for cpu-1:6ef6c041-2b0d-4bc3-8662-5d679a77a86d: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 6ef6c041-2b0d-4bc3-8662-5d679a77a86d could not be found. [ 496.320204] env[63021]: INFO nova.compute.resource_tracker [None req-1d0411b1-5d94-4c30-8c63-f930ce6d26d8 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d [ 497.840757] env[63021]: DEBUG nova.compute.resource_tracker [None req-1d0411b1-5d94-4c30-8c63-f930ce6d26d8 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=63021) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 497.841131] env[63021]: DEBUG nova.compute.resource_tracker [None req-1d0411b1-5d94-4c30-8c63-f930ce6d26d8 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=63021) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 497.996174] env[63021]: INFO nova.scheduler.client.report [None req-1d0411b1-5d94-4c30-8c63-f930ce6d26d8 None None] [req-1f9c528e-c9df-4c4c-87e0-06610268a9e2] Created resource provider record via placement API for resource provider with UUID 6ef6c041-2b0d-4bc3-8662-5d679a77a86d and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 498.011875] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb56e62d-68e6-4eec-8965-baad8cf62bda {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 498.019580] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bbb5583-edb6-4321-b944-6019c6885c5d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 498.049923] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2cda618-f589-4eba-a2c1-5e4cd90c2c19 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 498.056938] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0db0e05-db9a-4987-8882-725f95214353 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 498.069606] env[63021]: DEBUG nova.compute.provider_tree [None req-1d0411b1-5d94-4c30-8c63-f930ce6d26d8 None None] Updating inventory in ProviderTree for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 498.623107] env[63021]: DEBUG nova.scheduler.client.report [None req-1d0411b1-5d94-4c30-8c63-f930ce6d26d8 None None] Updated inventory for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 498.623343] env[63021]: DEBUG nova.compute.provider_tree [None req-1d0411b1-5d94-4c30-8c63-f930ce6d26d8 None None] Updating resource provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d generation from 0 to 1 during operation: update_inventory {{(pid=63021) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 498.623481] env[63021]: DEBUG nova.compute.provider_tree [None req-1d0411b1-5d94-4c30-8c63-f930ce6d26d8 None None] Updating inventory in ProviderTree for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 498.675436] env[63021]: DEBUG nova.compute.provider_tree [None req-1d0411b1-5d94-4c30-8c63-f930ce6d26d8 None None] Updating resource provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d generation from 1 to 2 during operation: update_traits {{(pid=63021) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 499.180398] env[63021]: DEBUG nova.compute.resource_tracker [None req-1d0411b1-5d94-4c30-8c63-f930ce6d26d8 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63021) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 499.180755] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1d0411b1-5d94-4c30-8c63-f930ce6d26d8 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.867s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 499.180755] env[63021]: DEBUG nova.service [None req-1d0411b1-5d94-4c30-8c63-f930ce6d26d8 None None] Creating RPC server for service compute {{(pid=63021) start /opt/stack/nova/nova/service.py:186}} [ 499.193509] env[63021]: DEBUG nova.service [None req-1d0411b1-5d94-4c30-8c63-f930ce6d26d8 None None] Join ServiceGroup membership for this service compute {{(pid=63021) start /opt/stack/nova/nova/service.py:203}} [ 499.193681] env[63021]: DEBUG nova.servicegroup.drivers.db [None req-1d0411b1-5d94-4c30-8c63-f930ce6d26d8 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=63021) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 535.195218] env[63021]: DEBUG oslo_concurrency.lockutils [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Acquiring lock "4abfadf8-b4cc-45ef-94b2-fcc6db6afd71" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.195218] env[63021]: DEBUG oslo_concurrency.lockutils [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Lock "4abfadf8-b4cc-45ef-94b2-fcc6db6afd71" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.197255] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._sync_power_states {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 535.700120] env[63021]: DEBUG nova.compute.manager [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 535.708389] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Getting list of instances from cluster (obj){ [ 535.708389] env[63021]: value = "domain-c8" [ 535.708389] env[63021]: _type = "ClusterComputeResource" [ 535.708389] env[63021]: } {{(pid=63021) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 535.709914] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be663108-1739-44d7-8070-6740ac0f0fc0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.721436] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Got total of 0 instances {{(pid=63021) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 535.721929] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 535.722382] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Getting list of instances from cluster (obj){ [ 535.722382] env[63021]: value = "domain-c8" [ 535.722382] env[63021]: _type = "ClusterComputeResource" [ 535.722382] env[63021]: } {{(pid=63021) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 535.723647] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a1d9a47-2c71-4d5a-93fe-cd7d068be7d3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.733227] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Got total of 0 instances {{(pid=63021) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 536.259751] env[63021]: DEBUG oslo_concurrency.lockutils [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.259751] env[63021]: DEBUG oslo_concurrency.lockutils [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.263754] env[63021]: INFO nova.compute.claims [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 536.453697] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Acquiring lock "31b903cb-64b8-46f4-b96f-1d3337ad776d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.453960] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Lock "31b903cb-64b8-46f4-b96f-1d3337ad776d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.960247] env[63021]: DEBUG nova.compute.manager [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 537.346803] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6b916f1-de3a-461c-bb33-301134e9986e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.355537] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adfebbbb-1b84-4dde-82c0-3540494503fd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.393028] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2014a125-5195-46a6-974a-c51c6cf650de {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.400219] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc9dcf33-9323-4948-823d-2db50fef5232 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.415582] env[63021]: DEBUG nova.compute.provider_tree [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 537.489273] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.549282] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Acquiring lock "d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.550551] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Lock "d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.920438] env[63021]: DEBUG nova.scheduler.client.report [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 538.060209] env[63021]: DEBUG nova.compute.manager [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 538.427095] env[63021]: DEBUG oslo_concurrency.lockutils [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.169s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 538.427655] env[63021]: DEBUG nova.compute.manager [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 538.431068] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.942s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.434255] env[63021]: INFO nova.compute.claims [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 538.591756] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.935175] env[63021]: DEBUG nova.compute.utils [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 538.936220] env[63021]: DEBUG nova.compute.manager [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 538.936453] env[63021]: DEBUG nova.network.neutron [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 539.446441] env[63021]: DEBUG nova.compute.manager [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 539.520028] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ab3dd45-a80d-49cb-b4e6-032af1caddf6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.527594] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-446a59f2-8c9a-4bdc-a891-4cb2c12bbdf2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.559863] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0006726d-0130-474c-b7e5-34f68ceb5d8d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.567651] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd67e4a0-c272-4a46-8e18-dc8cf17a6ac8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.582049] env[63021]: DEBUG nova.compute.provider_tree [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 540.078932] env[63021]: DEBUG nova.policy [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c468d6ce6b74478eb5adef00f6f05bb7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4df48a4180a943b5911f9d287b86f016', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 540.085116] env[63021]: DEBUG nova.scheduler.client.report [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 540.459831] env[63021]: DEBUG nova.compute.manager [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 540.497417] env[63021]: DEBUG nova.virt.hardware [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 540.497417] env[63021]: DEBUG nova.virt.hardware [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 540.497417] env[63021]: DEBUG nova.virt.hardware [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 540.497735] env[63021]: DEBUG nova.virt.hardware [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 540.497735] env[63021]: DEBUG nova.virt.hardware [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 540.497735] env[63021]: DEBUG nova.virt.hardware [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 540.497735] env[63021]: DEBUG nova.virt.hardware [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 540.497735] env[63021]: DEBUG nova.virt.hardware [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 540.497938] env[63021]: DEBUG nova.virt.hardware [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 540.497938] env[63021]: DEBUG nova.virt.hardware [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 540.497938] env[63021]: DEBUG nova.virt.hardware [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 540.498687] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cc1e40a-e59f-4de4-9059-9a54da8ba5b7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.509270] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95f19265-7241-43fb-918a-3ec53617f4a3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.528478] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beb04b07-9a76-4ba0-a1fa-ce81d23d349a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.565885] env[63021]: DEBUG nova.network.neutron [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Successfully created port: 320ec193-02e6-4115-9960-e43e08bc2a39 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 540.593103] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.162s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 540.593714] env[63021]: DEBUG nova.compute.manager [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 540.596313] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.005s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.598020] env[63021]: INFO nova.compute.claims [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 541.102927] env[63021]: DEBUG nova.compute.utils [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 541.104448] env[63021]: DEBUG nova.compute.manager [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 541.104570] env[63021]: DEBUG nova.network.neutron [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 541.206749] env[63021]: DEBUG nova.policy [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '44d816a067324b128c8795f068ff2b7a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '78bbd2fad179498d926ace8d0e6c94cc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 541.217658] env[63021]: DEBUG oslo_concurrency.lockutils [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Acquiring lock "d629c773-3eb1-4f8d-9888-8bbf7565acea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.218193] env[63021]: DEBUG oslo_concurrency.lockutils [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Lock "d629c773-3eb1-4f8d-9888-8bbf7565acea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.615932] env[63021]: DEBUG nova.compute.manager [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 541.723638] env[63021]: DEBUG nova.compute.manager [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 541.749078] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db3a0724-c34d-4f32-bff0-7bf118c55bf4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.764810] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b109513c-cfe8-4479-b819-b0c1e6c14f60 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.802127] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aab4235-cbaf-4972-a9c3-e7517bb15d0b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.811840] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95cdc9e4-477a-47ed-b7d5-1370dd94c6b7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.827740] env[63021]: DEBUG nova.compute.provider_tree [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 542.253323] env[63021]: DEBUG oslo_concurrency.lockutils [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.334397] env[63021]: DEBUG nova.scheduler.client.report [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 542.439438] env[63021]: DEBUG oslo_concurrency.lockutils [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Acquiring lock "0b443010-4dcf-4a7a-aec3-2d50bd50acc3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.440265] env[63021]: DEBUG oslo_concurrency.lockutils [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Lock "0b443010-4dcf-4a7a-aec3-2d50bd50acc3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.459702] env[63021]: DEBUG nova.network.neutron [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Successfully created port: fbbedcd4-9539-41e5-9059-34e6281b2574 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 542.629770] env[63021]: DEBUG nova.compute.manager [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 542.664300] env[63021]: DEBUG nova.virt.hardware [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 542.664478] env[63021]: DEBUG nova.virt.hardware [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 542.664626] env[63021]: DEBUG nova.virt.hardware [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 542.664795] env[63021]: DEBUG nova.virt.hardware [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 542.664925] env[63021]: DEBUG nova.virt.hardware [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 542.665166] env[63021]: DEBUG nova.virt.hardware [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 542.666070] env[63021]: DEBUG nova.virt.hardware [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 542.666070] env[63021]: DEBUG nova.virt.hardware [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 542.666070] env[63021]: DEBUG nova.virt.hardware [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 542.666070] env[63021]: DEBUG nova.virt.hardware [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 542.666070] env[63021]: DEBUG nova.virt.hardware [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 542.666826] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b875db1-2128-436f-83a5-9b070dbfa35b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.678305] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0573f23-deb1-4afa-b777-f4fa356b8f36 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.842964] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.244s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 542.842964] env[63021]: DEBUG nova.compute.manager [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 542.845277] env[63021]: DEBUG oslo_concurrency.lockutils [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.592s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.847196] env[63021]: INFO nova.compute.claims [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 542.942989] env[63021]: DEBUG nova.compute.manager [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 543.108527] env[63021]: DEBUG oslo_concurrency.lockutils [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Acquiring lock "7d98e2e4-33bc-4e6b-af6f-31c18f34b9af" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.108725] env[63021]: DEBUG oslo_concurrency.lockutils [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Lock "7d98e2e4-33bc-4e6b-af6f-31c18f34b9af" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.358134] env[63021]: DEBUG nova.compute.utils [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 543.367645] env[63021]: DEBUG nova.compute.manager [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 543.367645] env[63021]: DEBUG nova.network.neutron [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 543.475670] env[63021]: DEBUG oslo_concurrency.lockutils [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.553165] env[63021]: DEBUG nova.policy [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e063a962e064461994036bd68d51be76', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3275d10f82e94ecda609f9c7774146f5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 543.613270] env[63021]: DEBUG nova.compute.manager [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 543.870135] env[63021]: DEBUG nova.compute.manager [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 544.013806] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4076d1af-7792-4660-93fd-ed72027b4cc5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.023833] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dc11e44-2bbf-44a4-b0bf-2f24b3c7dacd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.062197] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1ad49a4-8b98-47b6-a41b-eef136d5c276 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.073868] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b077110-547f-469c-a146-428f4347a723 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.097609] env[63021]: DEBUG nova.compute.provider_tree [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 544.108158] env[63021]: ERROR nova.compute.manager [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 320ec193-02e6-4115-9960-e43e08bc2a39, please check neutron logs for more information. [ 544.108158] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 544.108158] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 544.108158] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 544.108158] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 544.108158] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 544.108158] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 544.108158] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 544.108158] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.108158] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 544.108158] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.108158] env[63021]: ERROR nova.compute.manager raise self.value [ 544.108158] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 544.108158] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 544.108158] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.108158] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 544.108662] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.108662] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 544.108662] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 320ec193-02e6-4115-9960-e43e08bc2a39, please check neutron logs for more information. [ 544.108662] env[63021]: ERROR nova.compute.manager [ 544.109082] env[63021]: Traceback (most recent call last): [ 544.109183] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 544.109183] env[63021]: listener.cb(fileno) [ 544.109252] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 544.109252] env[63021]: result = function(*args, **kwargs) [ 544.109332] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 544.109332] env[63021]: return func(*args, **kwargs) [ 544.110271] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 544.110271] env[63021]: raise e [ 544.110271] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 544.110271] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 544.110271] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 544.110271] env[63021]: created_port_ids = self._update_ports_for_instance( [ 544.110271] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 544.110271] env[63021]: with excutils.save_and_reraise_exception(): [ 544.110271] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.110271] env[63021]: self.force_reraise() [ 544.110271] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.110271] env[63021]: raise self.value [ 544.110271] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 544.110271] env[63021]: updated_port = self._update_port( [ 544.110271] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.110271] env[63021]: _ensure_no_port_binding_failure(port) [ 544.110271] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.110271] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 544.110271] env[63021]: nova.exception.PortBindingFailed: Binding failed for port 320ec193-02e6-4115-9960-e43e08bc2a39, please check neutron logs for more information. [ 544.110271] env[63021]: Removing descriptor: 15 [ 544.111053] env[63021]: ERROR nova.compute.manager [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 320ec193-02e6-4115-9960-e43e08bc2a39, please check neutron logs for more information. [ 544.111053] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Traceback (most recent call last): [ 544.111053] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 544.111053] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] yield resources [ 544.111053] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 544.111053] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] self.driver.spawn(context, instance, image_meta, [ 544.111053] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 544.111053] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] self._vmops.spawn(context, instance, image_meta, injected_files, [ 544.111053] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 544.111053] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] vm_ref = self.build_virtual_machine(instance, [ 544.111053] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 544.111347] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] vif_infos = vmwarevif.get_vif_info(self._session, [ 544.111347] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 544.111347] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] for vif in network_info: [ 544.111347] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 544.111347] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] return self._sync_wrapper(fn, *args, **kwargs) [ 544.111347] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 544.111347] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] self.wait() [ 544.111347] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 544.111347] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] self[:] = self._gt.wait() [ 544.111347] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 544.111347] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] return self._exit_event.wait() [ 544.111347] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 544.111347] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] result = hub.switch() [ 544.111708] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 544.111708] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] return self.greenlet.switch() [ 544.111708] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 544.111708] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] result = function(*args, **kwargs) [ 544.111708] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 544.111708] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] return func(*args, **kwargs) [ 544.111708] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 544.111708] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] raise e [ 544.111708] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 544.111708] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] nwinfo = self.network_api.allocate_for_instance( [ 544.111708] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 544.111708] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] created_port_ids = self._update_ports_for_instance( [ 544.111708] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 544.112051] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] with excutils.save_and_reraise_exception(): [ 544.112051] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.112051] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] self.force_reraise() [ 544.112051] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.112051] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] raise self.value [ 544.112051] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 544.112051] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] updated_port = self._update_port( [ 544.112051] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.112051] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] _ensure_no_port_binding_failure(port) [ 544.112051] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.112051] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] raise exception.PortBindingFailed(port_id=port['id']) [ 544.112051] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] nova.exception.PortBindingFailed: Binding failed for port 320ec193-02e6-4115-9960-e43e08bc2a39, please check neutron logs for more information. [ 544.112051] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] [ 544.112442] env[63021]: INFO nova.compute.manager [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Terminating instance [ 544.113618] env[63021]: DEBUG oslo_concurrency.lockutils [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Acquiring lock "refresh_cache-4abfadf8-b4cc-45ef-94b2-fcc6db6afd71" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 544.116998] env[63021]: DEBUG oslo_concurrency.lockutils [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Acquired lock "refresh_cache-4abfadf8-b4cc-45ef-94b2-fcc6db6afd71" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.116998] env[63021]: DEBUG nova.network.neutron [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 544.134764] env[63021]: DEBUG oslo_concurrency.lockutils [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.601479] env[63021]: DEBUG nova.scheduler.client.report [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 544.674318] env[63021]: DEBUG nova.network.neutron [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 544.852826] env[63021]: DEBUG nova.network.neutron [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.886934] env[63021]: DEBUG nova.compute.manager [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 544.920367] env[63021]: DEBUG nova.virt.hardware [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 544.920367] env[63021]: DEBUG nova.virt.hardware [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 544.920367] env[63021]: DEBUG nova.virt.hardware [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 544.920367] env[63021]: DEBUG nova.virt.hardware [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 544.920585] env[63021]: DEBUG nova.virt.hardware [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 544.920585] env[63021]: DEBUG nova.virt.hardware [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 544.920585] env[63021]: DEBUG nova.virt.hardware [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 544.920585] env[63021]: DEBUG nova.virt.hardware [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 544.920585] env[63021]: DEBUG nova.virt.hardware [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 544.920745] env[63021]: DEBUG nova.virt.hardware [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 544.920745] env[63021]: DEBUG nova.virt.hardware [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 544.921147] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f981574a-18e2-4c3b-8f00-460cb74ee049 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.932807] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53a1124d-ebb8-47b2-b0d7-9014cbd4da4a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.110103] env[63021]: DEBUG oslo_concurrency.lockutils [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.264s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.111988] env[63021]: DEBUG nova.compute.manager [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 545.121028] env[63021]: DEBUG oslo_concurrency.lockutils [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.640s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.121028] env[63021]: INFO nova.compute.claims [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 545.126285] env[63021]: DEBUG nova.network.neutron [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Successfully created port: a1a43b46-a34b-4ebe-befa-e72aa56aae02 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 545.196767] env[63021]: ERROR nova.compute.manager [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fbbedcd4-9539-41e5-9059-34e6281b2574, please check neutron logs for more information. [ 545.196767] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 545.196767] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 545.196767] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 545.196767] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 545.196767] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 545.196767] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 545.196767] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 545.196767] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.196767] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 545.196767] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.196767] env[63021]: ERROR nova.compute.manager raise self.value [ 545.196767] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 545.196767] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 545.196767] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.196767] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 545.197230] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.197230] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 545.197230] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fbbedcd4-9539-41e5-9059-34e6281b2574, please check neutron logs for more information. [ 545.197230] env[63021]: ERROR nova.compute.manager [ 545.197230] env[63021]: Traceback (most recent call last): [ 545.197230] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 545.197230] env[63021]: listener.cb(fileno) [ 545.197230] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 545.197230] env[63021]: result = function(*args, **kwargs) [ 545.197230] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 545.197230] env[63021]: return func(*args, **kwargs) [ 545.197230] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 545.197230] env[63021]: raise e [ 545.197230] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 545.197230] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 545.197230] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 545.197230] env[63021]: created_port_ids = self._update_ports_for_instance( [ 545.197230] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 545.197230] env[63021]: with excutils.save_and_reraise_exception(): [ 545.197230] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.197230] env[63021]: self.force_reraise() [ 545.197230] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.197230] env[63021]: raise self.value [ 545.197230] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 545.197230] env[63021]: updated_port = self._update_port( [ 545.197230] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.197230] env[63021]: _ensure_no_port_binding_failure(port) [ 545.197230] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.197230] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 545.197960] env[63021]: nova.exception.PortBindingFailed: Binding failed for port fbbedcd4-9539-41e5-9059-34e6281b2574, please check neutron logs for more information. [ 545.197960] env[63021]: Removing descriptor: 16 [ 545.197960] env[63021]: ERROR nova.compute.manager [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fbbedcd4-9539-41e5-9059-34e6281b2574, please check neutron logs for more information. [ 545.197960] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Traceback (most recent call last): [ 545.197960] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 545.197960] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] yield resources [ 545.197960] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 545.197960] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] self.driver.spawn(context, instance, image_meta, [ 545.197960] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 545.197960] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 545.197960] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 545.197960] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] vm_ref = self.build_virtual_machine(instance, [ 545.198281] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 545.198281] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] vif_infos = vmwarevif.get_vif_info(self._session, [ 545.198281] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 545.198281] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] for vif in network_info: [ 545.198281] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 545.198281] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] return self._sync_wrapper(fn, *args, **kwargs) [ 545.198281] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 545.198281] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] self.wait() [ 545.198281] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 545.198281] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] self[:] = self._gt.wait() [ 545.198281] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 545.198281] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] return self._exit_event.wait() [ 545.198281] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 545.201500] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] result = hub.switch() [ 545.201500] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 545.201500] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] return self.greenlet.switch() [ 545.201500] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 545.201500] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] result = function(*args, **kwargs) [ 545.201500] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 545.201500] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] return func(*args, **kwargs) [ 545.201500] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 545.201500] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] raise e [ 545.201500] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 545.201500] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] nwinfo = self.network_api.allocate_for_instance( [ 545.201500] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 545.201500] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] created_port_ids = self._update_ports_for_instance( [ 545.201978] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 545.201978] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] with excutils.save_and_reraise_exception(): [ 545.201978] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.201978] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] self.force_reraise() [ 545.201978] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.201978] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] raise self.value [ 545.201978] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 545.201978] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] updated_port = self._update_port( [ 545.201978] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.201978] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] _ensure_no_port_binding_failure(port) [ 545.201978] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.201978] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] raise exception.PortBindingFailed(port_id=port['id']) [ 545.202340] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] nova.exception.PortBindingFailed: Binding failed for port fbbedcd4-9539-41e5-9059-34e6281b2574, please check neutron logs for more information. [ 545.202340] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] [ 545.202340] env[63021]: INFO nova.compute.manager [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Terminating instance [ 545.202340] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Acquiring lock "refresh_cache-31b903cb-64b8-46f4-b96f-1d3337ad776d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.202494] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Acquired lock "refresh_cache-31b903cb-64b8-46f4-b96f-1d3337ad776d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 545.202619] env[63021]: DEBUG nova.network.neutron [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 545.358697] env[63021]: DEBUG oslo_concurrency.lockutils [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Releasing lock "refresh_cache-4abfadf8-b4cc-45ef-94b2-fcc6db6afd71" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 545.359263] env[63021]: DEBUG nova.compute.manager [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 545.359352] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 545.359774] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2d315ad7-6c23-46a4-8674-038b40a9ac46 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.372395] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d12a7ab-bf43-4215-84d0-d052be1b1fe6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.400912] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71 could not be found. [ 545.401163] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 545.402247] env[63021]: INFO nova.compute.manager [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Took 0.04 seconds to destroy the instance on the hypervisor. [ 545.402247] env[63021]: DEBUG oslo.service.loopingcall [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 545.402247] env[63021]: DEBUG nova.compute.manager [-] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 545.402247] env[63021]: DEBUG nova.network.neutron [-] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 545.437992] env[63021]: DEBUG nova.network.neutron [-] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 545.479790] env[63021]: DEBUG nova.compute.manager [req-032fe28d-d22b-4d34-8d8e-1cf08442994d req-e3caf4b1-7c79-4eaa-af7d-7b5f88f50bd6 service nova] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Received event network-changed-320ec193-02e6-4115-9960-e43e08bc2a39 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 545.479790] env[63021]: DEBUG nova.compute.manager [req-032fe28d-d22b-4d34-8d8e-1cf08442994d req-e3caf4b1-7c79-4eaa-af7d-7b5f88f50bd6 service nova] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Refreshing instance network info cache due to event network-changed-320ec193-02e6-4115-9960-e43e08bc2a39. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 545.480188] env[63021]: DEBUG oslo_concurrency.lockutils [req-032fe28d-d22b-4d34-8d8e-1cf08442994d req-e3caf4b1-7c79-4eaa-af7d-7b5f88f50bd6 service nova] Acquiring lock "refresh_cache-4abfadf8-b4cc-45ef-94b2-fcc6db6afd71" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.480188] env[63021]: DEBUG oslo_concurrency.lockutils [req-032fe28d-d22b-4d34-8d8e-1cf08442994d req-e3caf4b1-7c79-4eaa-af7d-7b5f88f50bd6 service nova] Acquired lock "refresh_cache-4abfadf8-b4cc-45ef-94b2-fcc6db6afd71" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 545.480249] env[63021]: DEBUG nova.network.neutron [req-032fe28d-d22b-4d34-8d8e-1cf08442994d req-e3caf4b1-7c79-4eaa-af7d-7b5f88f50bd6 service nova] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Refreshing network info cache for port 320ec193-02e6-4115-9960-e43e08bc2a39 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 545.623737] env[63021]: DEBUG nova.compute.utils [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 545.635698] env[63021]: DEBUG nova.compute.manager [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 545.635698] env[63021]: DEBUG nova.network.neutron [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 545.807485] env[63021]: DEBUG nova.network.neutron [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 545.944408] env[63021]: DEBUG nova.network.neutron [-] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.000797] env[63021]: DEBUG nova.policy [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1e98df5d168d47c185615e7ac5bb6b2e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1175ae6d5d314127bee4edd4d5308fb0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 546.055050] env[63021]: DEBUG nova.network.neutron [req-032fe28d-d22b-4d34-8d8e-1cf08442994d req-e3caf4b1-7c79-4eaa-af7d-7b5f88f50bd6 service nova] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 546.104194] env[63021]: DEBUG nova.network.neutron [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.130723] env[63021]: DEBUG nova.compute.manager [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 546.280642] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6daa07aa-4197-451b-8103-64c4fa196567 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.286294] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbe97551-4049-44f5-9045-7382cbb08840 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.319038] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c93b3ca6-fb0f-4420-b21f-d84ad6e89c2b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.326500] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1a49b11-50f8-47c4-855b-86d8140688be {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.339926] env[63021]: DEBUG nova.compute.provider_tree [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 546.349258] env[63021]: DEBUG nova.network.neutron [req-032fe28d-d22b-4d34-8d8e-1cf08442994d req-e3caf4b1-7c79-4eaa-af7d-7b5f88f50bd6 service nova] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.446233] env[63021]: INFO nova.compute.manager [-] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Took 1.04 seconds to deallocate network for instance. [ 546.449349] env[63021]: DEBUG nova.compute.claims [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 546.449532] env[63021]: DEBUG oslo_concurrency.lockutils [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.606872] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Releasing lock "refresh_cache-31b903cb-64b8-46f4-b96f-1d3337ad776d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 546.607419] env[63021]: DEBUG nova.compute.manager [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 546.607614] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 546.607910] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bb98842a-9110-4917-b067-cda9025ea9ff {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.617609] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84639cd5-5fc4-4d3b-9496-18b5095e74b9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.643385] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 31b903cb-64b8-46f4-b96f-1d3337ad776d could not be found. [ 546.643647] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 546.644824] env[63021]: INFO nova.compute.manager [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 546.644824] env[63021]: DEBUG oslo.service.loopingcall [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 546.645739] env[63021]: DEBUG nova.compute.manager [-] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 546.646037] env[63021]: DEBUG nova.network.neutron [-] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 546.671768] env[63021]: DEBUG nova.network.neutron [-] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 546.811579] env[63021]: DEBUG nova.network.neutron [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Successfully created port: 80b3a05d-0a47-4bb2-ae3b-41543c11af6c {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 546.843997] env[63021]: DEBUG nova.scheduler.client.report [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 546.854131] env[63021]: DEBUG oslo_concurrency.lockutils [req-032fe28d-d22b-4d34-8d8e-1cf08442994d req-e3caf4b1-7c79-4eaa-af7d-7b5f88f50bd6 service nova] Releasing lock "refresh_cache-4abfadf8-b4cc-45ef-94b2-fcc6db6afd71" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.148363] env[63021]: DEBUG nova.compute.manager [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 547.177518] env[63021]: DEBUG nova.network.neutron [-] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.188408] env[63021]: DEBUG nova.virt.hardware [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 547.188757] env[63021]: DEBUG nova.virt.hardware [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 547.188896] env[63021]: DEBUG nova.virt.hardware [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 547.189328] env[63021]: DEBUG nova.virt.hardware [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 547.189328] env[63021]: DEBUG nova.virt.hardware [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 547.189441] env[63021]: DEBUG nova.virt.hardware [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 547.189712] env[63021]: DEBUG nova.virt.hardware [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 547.189712] env[63021]: DEBUG nova.virt.hardware [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 547.189976] env[63021]: DEBUG nova.virt.hardware [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 547.190037] env[63021]: DEBUG nova.virt.hardware [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 547.190561] env[63021]: DEBUG nova.virt.hardware [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 547.191269] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be11d3cf-8bb5-49af-8589-1a3835c43ab2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.201712] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85ecbf79-918a-4df6-9410-e211b32ecdc3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.352779] env[63021]: DEBUG oslo_concurrency.lockutils [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.236s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 547.352779] env[63021]: DEBUG nova.compute.manager [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 547.358942] env[63021]: DEBUG oslo_concurrency.lockutils [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.224s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.361334] env[63021]: INFO nova.compute.claims [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 547.613583] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 547.613583] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 547.613583] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Starting heal instance info cache {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 547.613583] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Rebuilding the list of instances to heal {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 547.679579] env[63021]: INFO nova.compute.manager [-] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Took 1.03 seconds to deallocate network for instance. [ 547.683492] env[63021]: DEBUG nova.compute.claims [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 547.684233] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.757352] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Acquiring lock "ea199990-9638-4e94-87fe-bd4b7f020595" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.757611] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Lock "ea199990-9638-4e94-87fe-bd4b7f020595" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.860918] env[63021]: DEBUG nova.compute.utils [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 547.862523] env[63021]: DEBUG nova.compute.manager [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 547.862817] env[63021]: DEBUG nova.network.neutron [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 548.099744] env[63021]: DEBUG nova.policy [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dfd0a7c7e22941cfb8b21acd5bb3a5b0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6bb23a9fa9a847839e414c27d1def665', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 548.120338] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Skipping network cache update for instance because it is Building. {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 548.120492] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Skipping network cache update for instance because it is Building. {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 548.120644] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Skipping network cache update for instance because it is Building. {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 548.120780] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Skipping network cache update for instance because it is Building. {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 548.120899] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Skipping network cache update for instance because it is Building. {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 548.121027] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Skipping network cache update for instance because it is Building. {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 548.121152] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Didn't find any instances for network info cache update. {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 548.121347] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 548.121578] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 548.122039] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 548.122039] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 548.122151] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 548.122347] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 548.122505] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63021) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 548.122643] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 548.260144] env[63021]: DEBUG nova.compute.manager [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 548.365626] env[63021]: DEBUG nova.compute.manager [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 548.515090] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bf96a58-6a1c-48a0-8a76-267919793d2e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.525135] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb0bd97f-4196-44cd-beeb-80fc9227f3bd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.563096] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a931a6fc-6773-4812-80ec-1e9a7c92fb54 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.574648] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39f06949-c1ee-4626-95d6-9f37b09fd1c7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.588246] env[63021]: DEBUG nova.compute.provider_tree [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 548.591679] env[63021]: DEBUG nova.compute.manager [req-03c77ec2-13a3-490a-a04c-90a6a44fab46 req-ec7f7f84-2072-4904-81d1-cbbc68b5f16b service nova] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Received event network-vif-deleted-320ec193-02e6-4115-9960-e43e08bc2a39 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 548.591679] env[63021]: DEBUG nova.compute.manager [req-03c77ec2-13a3-490a-a04c-90a6a44fab46 req-ec7f7f84-2072-4904-81d1-cbbc68b5f16b service nova] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Received event network-changed-fbbedcd4-9539-41e5-9059-34e6281b2574 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 548.591859] env[63021]: DEBUG nova.compute.manager [req-03c77ec2-13a3-490a-a04c-90a6a44fab46 req-ec7f7f84-2072-4904-81d1-cbbc68b5f16b service nova] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Refreshing instance network info cache due to event network-changed-fbbedcd4-9539-41e5-9059-34e6281b2574. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 548.591923] env[63021]: DEBUG oslo_concurrency.lockutils [req-03c77ec2-13a3-490a-a04c-90a6a44fab46 req-ec7f7f84-2072-4904-81d1-cbbc68b5f16b service nova] Acquiring lock "refresh_cache-31b903cb-64b8-46f4-b96f-1d3337ad776d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.593311] env[63021]: DEBUG oslo_concurrency.lockutils [req-03c77ec2-13a3-490a-a04c-90a6a44fab46 req-ec7f7f84-2072-4904-81d1-cbbc68b5f16b service nova] Acquired lock "refresh_cache-31b903cb-64b8-46f4-b96f-1d3337ad776d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.593311] env[63021]: DEBUG nova.network.neutron [req-03c77ec2-13a3-490a-a04c-90a6a44fab46 req-ec7f7f84-2072-4904-81d1-cbbc68b5f16b service nova] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Refreshing network info cache for port fbbedcd4-9539-41e5-9059-34e6281b2574 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 548.625849] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.788239] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.094256] env[63021]: DEBUG nova.scheduler.client.report [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 549.116369] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Acquiring lock "0950cbd3-0708-4994-8d2d-053aecc06cdf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.116369] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Lock "0950cbd3-0708-4994-8d2d-053aecc06cdf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.142195] env[63021]: DEBUG nova.network.neutron [req-03c77ec2-13a3-490a-a04c-90a6a44fab46 req-ec7f7f84-2072-4904-81d1-cbbc68b5f16b service nova] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 549.379079] env[63021]: DEBUG nova.compute.manager [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 549.382384] env[63021]: DEBUG nova.network.neutron [req-03c77ec2-13a3-490a-a04c-90a6a44fab46 req-ec7f7f84-2072-4904-81d1-cbbc68b5f16b service nova] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.414491] env[63021]: DEBUG nova.virt.hardware [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 549.415402] env[63021]: DEBUG nova.virt.hardware [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 549.415727] env[63021]: DEBUG nova.virt.hardware [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 549.415956] env[63021]: DEBUG nova.virt.hardware [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 549.416127] env[63021]: DEBUG nova.virt.hardware [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 549.416274] env[63021]: DEBUG nova.virt.hardware [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 549.416477] env[63021]: DEBUG nova.virt.hardware [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 549.416626] env[63021]: DEBUG nova.virt.hardware [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 549.417185] env[63021]: DEBUG nova.virt.hardware [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 549.417185] env[63021]: DEBUG nova.virt.hardware [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 549.417185] env[63021]: DEBUG nova.virt.hardware [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 549.421278] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc73850a-7492-4280-87c6-15553935c1af {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.430491] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-517270c9-1f1d-4f0f-9d1a-a80bb1d077a9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.599323] env[63021]: DEBUG nova.network.neutron [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Successfully created port: b1544ac9-c61e-4f0a-a311-99a23b19e457 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 549.610715] env[63021]: DEBUG oslo_concurrency.lockutils [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.247s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 549.610715] env[63021]: DEBUG nova.compute.manager [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 549.613520] env[63021]: DEBUG oslo_concurrency.lockutils [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 3.160s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.625562] env[63021]: DEBUG nova.compute.manager [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 549.885594] env[63021]: DEBUG oslo_concurrency.lockutils [req-03c77ec2-13a3-490a-a04c-90a6a44fab46 req-ec7f7f84-2072-4904-81d1-cbbc68b5f16b service nova] Releasing lock "refresh_cache-31b903cb-64b8-46f4-b96f-1d3337ad776d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.885876] env[63021]: DEBUG nova.compute.manager [req-03c77ec2-13a3-490a-a04c-90a6a44fab46 req-ec7f7f84-2072-4904-81d1-cbbc68b5f16b service nova] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Received event network-vif-deleted-fbbedcd4-9539-41e5-9059-34e6281b2574 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 550.122748] env[63021]: DEBUG nova.compute.utils [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 550.132215] env[63021]: DEBUG nova.compute.manager [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 550.132215] env[63021]: DEBUG nova.network.neutron [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 550.168858] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.288787] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-248aa632-24fb-495d-8873-c3ddf1c5687c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.299638] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbb2a1f2-6e95-4701-b0b0-3318c709cfdc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.330867] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae6d1b16-ec54-4ba6-881b-98c3d065452c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.341025] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72f703d9-384e-4ab9-b628-028766e0c3bf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.358306] env[63021]: DEBUG nova.compute.provider_tree [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 550.468856] env[63021]: DEBUG nova.policy [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8206914e1f2244c2a8a713f80435cee9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '049f7455a1274885a71766f01aaf042d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 550.632729] env[63021]: DEBUG nova.compute.manager [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 550.861891] env[63021]: DEBUG nova.scheduler.client.report [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 551.368487] env[63021]: DEBUG oslo_concurrency.lockutils [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.759s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.371432] env[63021]: ERROR nova.compute.manager [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 320ec193-02e6-4115-9960-e43e08bc2a39, please check neutron logs for more information. [ 551.371432] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Traceback (most recent call last): [ 551.371432] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 551.371432] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] self.driver.spawn(context, instance, image_meta, [ 551.371432] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 551.371432] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] self._vmops.spawn(context, instance, image_meta, injected_files, [ 551.371432] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 551.371432] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] vm_ref = self.build_virtual_machine(instance, [ 551.371432] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 551.371432] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] vif_infos = vmwarevif.get_vif_info(self._session, [ 551.371432] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 551.372051] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] for vif in network_info: [ 551.372051] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 551.372051] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] return self._sync_wrapper(fn, *args, **kwargs) [ 551.372051] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 551.372051] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] self.wait() [ 551.372051] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 551.372051] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] self[:] = self._gt.wait() [ 551.372051] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 551.372051] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] return self._exit_event.wait() [ 551.372051] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 551.372051] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] result = hub.switch() [ 551.372051] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 551.372051] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] return self.greenlet.switch() [ 551.372402] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 551.372402] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] result = function(*args, **kwargs) [ 551.372402] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 551.372402] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] return func(*args, **kwargs) [ 551.372402] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 551.372402] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] raise e [ 551.372402] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.372402] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] nwinfo = self.network_api.allocate_for_instance( [ 551.372402] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 551.372402] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] created_port_ids = self._update_ports_for_instance( [ 551.372402] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 551.372402] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] with excutils.save_and_reraise_exception(): [ 551.372402] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.372719] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] self.force_reraise() [ 551.372719] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.372719] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] raise self.value [ 551.372719] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 551.372719] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] updated_port = self._update_port( [ 551.372719] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.372719] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] _ensure_no_port_binding_failure(port) [ 551.372719] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.372719] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] raise exception.PortBindingFailed(port_id=port['id']) [ 551.372719] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] nova.exception.PortBindingFailed: Binding failed for port 320ec193-02e6-4115-9960-e43e08bc2a39, please check neutron logs for more information. [ 551.372719] env[63021]: ERROR nova.compute.manager [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] [ 551.373054] env[63021]: DEBUG nova.compute.utils [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Binding failed for port 320ec193-02e6-4115-9960-e43e08bc2a39, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 551.373054] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 3.687s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.378608] env[63021]: DEBUG nova.compute.manager [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Build of instance 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71 was re-scheduled: Binding failed for port 320ec193-02e6-4115-9960-e43e08bc2a39, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 551.379095] env[63021]: DEBUG nova.compute.manager [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 551.379322] env[63021]: DEBUG oslo_concurrency.lockutils [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Acquiring lock "refresh_cache-4abfadf8-b4cc-45ef-94b2-fcc6db6afd71" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 551.379459] env[63021]: DEBUG oslo_concurrency.lockutils [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Acquired lock "refresh_cache-4abfadf8-b4cc-45ef-94b2-fcc6db6afd71" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 551.379610] env[63021]: DEBUG nova.network.neutron [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 551.582197] env[63021]: ERROR nova.compute.manager [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a1a43b46-a34b-4ebe-befa-e72aa56aae02, please check neutron logs for more information. [ 551.582197] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 551.582197] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.582197] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 551.582197] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 551.582197] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 551.582197] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 551.582197] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 551.582197] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.582197] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 551.582197] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.582197] env[63021]: ERROR nova.compute.manager raise self.value [ 551.582197] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 551.582197] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 551.582197] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.582197] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 551.583153] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.583153] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 551.583153] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a1a43b46-a34b-4ebe-befa-e72aa56aae02, please check neutron logs for more information. [ 551.583153] env[63021]: ERROR nova.compute.manager [ 551.583153] env[63021]: Traceback (most recent call last): [ 551.583153] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 551.583153] env[63021]: listener.cb(fileno) [ 551.583153] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 551.583153] env[63021]: result = function(*args, **kwargs) [ 551.583153] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 551.583153] env[63021]: return func(*args, **kwargs) [ 551.583153] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 551.583153] env[63021]: raise e [ 551.583153] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.583153] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 551.583153] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 551.583153] env[63021]: created_port_ids = self._update_ports_for_instance( [ 551.583153] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 551.583153] env[63021]: with excutils.save_and_reraise_exception(): [ 551.583153] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.583153] env[63021]: self.force_reraise() [ 551.583153] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.583153] env[63021]: raise self.value [ 551.583153] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 551.583153] env[63021]: updated_port = self._update_port( [ 551.583153] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.583153] env[63021]: _ensure_no_port_binding_failure(port) [ 551.583153] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.583153] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 551.583854] env[63021]: nova.exception.PortBindingFailed: Binding failed for port a1a43b46-a34b-4ebe-befa-e72aa56aae02, please check neutron logs for more information. [ 551.583854] env[63021]: Removing descriptor: 17 [ 551.583854] env[63021]: ERROR nova.compute.manager [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a1a43b46-a34b-4ebe-befa-e72aa56aae02, please check neutron logs for more information. [ 551.583854] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Traceback (most recent call last): [ 551.583854] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 551.583854] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] yield resources [ 551.583854] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 551.583854] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] self.driver.spawn(context, instance, image_meta, [ 551.583854] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 551.583854] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 551.583854] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 551.583854] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] vm_ref = self.build_virtual_machine(instance, [ 551.584172] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 551.584172] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] vif_infos = vmwarevif.get_vif_info(self._session, [ 551.584172] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 551.584172] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] for vif in network_info: [ 551.584172] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 551.584172] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] return self._sync_wrapper(fn, *args, **kwargs) [ 551.584172] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 551.584172] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] self.wait() [ 551.584172] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 551.584172] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] self[:] = self._gt.wait() [ 551.584172] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 551.584172] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] return self._exit_event.wait() [ 551.584172] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 551.584504] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] result = hub.switch() [ 551.584504] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 551.584504] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] return self.greenlet.switch() [ 551.584504] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 551.584504] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] result = function(*args, **kwargs) [ 551.584504] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 551.584504] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] return func(*args, **kwargs) [ 551.584504] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 551.584504] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] raise e [ 551.584504] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.584504] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] nwinfo = self.network_api.allocate_for_instance( [ 551.584504] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 551.584504] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] created_port_ids = self._update_ports_for_instance( [ 551.584852] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 551.584852] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] with excutils.save_and_reraise_exception(): [ 551.584852] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.584852] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] self.force_reraise() [ 551.584852] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.584852] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] raise self.value [ 551.584852] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 551.584852] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] updated_port = self._update_port( [ 551.584852] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.584852] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] _ensure_no_port_binding_failure(port) [ 551.584852] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.584852] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] raise exception.PortBindingFailed(port_id=port['id']) [ 551.585150] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] nova.exception.PortBindingFailed: Binding failed for port a1a43b46-a34b-4ebe-befa-e72aa56aae02, please check neutron logs for more information. [ 551.585150] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] [ 551.585150] env[63021]: INFO nova.compute.manager [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Terminating instance [ 551.587397] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Acquiring lock "refresh_cache-d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 551.587595] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Acquired lock "refresh_cache-d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 551.587796] env[63021]: DEBUG nova.network.neutron [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 551.592749] env[63021]: DEBUG oslo_concurrency.lockutils [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Acquiring lock "2201799a-2df0-4fba-bcce-1fd8dfa1b6b9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.593325] env[63021]: DEBUG oslo_concurrency.lockutils [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Lock "2201799a-2df0-4fba-bcce-1fd8dfa1b6b9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.646019] env[63021]: DEBUG nova.compute.manager [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 551.678780] env[63021]: DEBUG nova.virt.hardware [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 551.679073] env[63021]: DEBUG nova.virt.hardware [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 551.679232] env[63021]: DEBUG nova.virt.hardware [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 551.679405] env[63021]: DEBUG nova.virt.hardware [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 551.679545] env[63021]: DEBUG nova.virt.hardware [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 551.679687] env[63021]: DEBUG nova.virt.hardware [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 551.679890] env[63021]: DEBUG nova.virt.hardware [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 551.680086] env[63021]: DEBUG nova.virt.hardware [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 551.680215] env[63021]: DEBUG nova.virt.hardware [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 551.680370] env[63021]: DEBUG nova.virt.hardware [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 551.680536] env[63021]: DEBUG nova.virt.hardware [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 551.681985] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c181e8e8-66c1-40e5-bc8b-f7b991d7de38 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.691746] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fe9e9df-2a50-4857-8d0e-1be8d36b5e8d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.718031] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Acquiring lock "5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.718031] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Lock "5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.980456] env[63021]: DEBUG nova.network.neutron [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 552.098674] env[63021]: DEBUG nova.compute.manager [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 552.145990] env[63021]: DEBUG nova.compute.manager [req-e87021f4-bfb6-4c04-a402-85f1240308f9 req-94bbd5d2-36fb-47ae-ab5c-d82adfd3c22c service nova] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Received event network-changed-a1a43b46-a34b-4ebe-befa-e72aa56aae02 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 552.146637] env[63021]: DEBUG nova.compute.manager [req-e87021f4-bfb6-4c04-a402-85f1240308f9 req-94bbd5d2-36fb-47ae-ab5c-d82adfd3c22c service nova] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Refreshing instance network info cache due to event network-changed-a1a43b46-a34b-4ebe-befa-e72aa56aae02. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 552.146637] env[63021]: DEBUG oslo_concurrency.lockutils [req-e87021f4-bfb6-4c04-a402-85f1240308f9 req-94bbd5d2-36fb-47ae-ab5c-d82adfd3c22c service nova] Acquiring lock "refresh_cache-d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 552.152618] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b65af45-50d8-499e-9b13-433647a74a23 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.160441] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9e58ce7-e42b-438c-ba01-6b55c2fee741 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.195813] env[63021]: DEBUG nova.network.neutron [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 552.198309] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec3d62e8-b549-46ca-b512-cdda971ae68f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.206552] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-526305fb-c85b-4e1c-bf7d-fecbd04c4364 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.221641] env[63021]: DEBUG nova.compute.provider_tree [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 552.224054] env[63021]: DEBUG nova.compute.manager [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 552.336264] env[63021]: DEBUG nova.network.neutron [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.609539] env[63021]: DEBUG nova.network.neutron [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.631679] env[63021]: DEBUG oslo_concurrency.lockutils [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.727991] env[63021]: DEBUG nova.scheduler.client.report [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 552.752929] env[63021]: DEBUG nova.compute.manager [req-9fae82ff-5847-4f9c-a74a-22958fb3767c req-db9a066e-2c9c-4486-913c-3d7d1061f911 service nova] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Received event network-changed-80b3a05d-0a47-4bb2-ae3b-41543c11af6c {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 552.752929] env[63021]: DEBUG nova.compute.manager [req-9fae82ff-5847-4f9c-a74a-22958fb3767c req-db9a066e-2c9c-4486-913c-3d7d1061f911 service nova] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Refreshing instance network info cache due to event network-changed-80b3a05d-0a47-4bb2-ae3b-41543c11af6c. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 552.753113] env[63021]: DEBUG oslo_concurrency.lockutils [req-9fae82ff-5847-4f9c-a74a-22958fb3767c req-db9a066e-2c9c-4486-913c-3d7d1061f911 service nova] Acquiring lock "refresh_cache-d629c773-3eb1-4f8d-9888-8bbf7565acea" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 552.753113] env[63021]: DEBUG oslo_concurrency.lockutils [req-9fae82ff-5847-4f9c-a74a-22958fb3767c req-db9a066e-2c9c-4486-913c-3d7d1061f911 service nova] Acquired lock "refresh_cache-d629c773-3eb1-4f8d-9888-8bbf7565acea" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 552.753252] env[63021]: DEBUG nova.network.neutron [req-9fae82ff-5847-4f9c-a74a-22958fb3767c req-db9a066e-2c9c-4486-913c-3d7d1061f911 service nova] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Refreshing network info cache for port 80b3a05d-0a47-4bb2-ae3b-41543c11af6c {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 552.767696] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.818354] env[63021]: ERROR nova.compute.manager [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 80b3a05d-0a47-4bb2-ae3b-41543c11af6c, please check neutron logs for more information. [ 552.818354] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 552.818354] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 552.818354] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 552.818354] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 552.818354] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 552.818354] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 552.818354] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 552.818354] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 552.818354] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 552.818354] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 552.818354] env[63021]: ERROR nova.compute.manager raise self.value [ 552.818354] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 552.818354] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 552.818354] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 552.818354] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 552.819906] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 552.819906] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 552.819906] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 80b3a05d-0a47-4bb2-ae3b-41543c11af6c, please check neutron logs for more information. [ 552.819906] env[63021]: ERROR nova.compute.manager [ 552.819906] env[63021]: Traceback (most recent call last): [ 552.819906] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 552.819906] env[63021]: listener.cb(fileno) [ 552.819906] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 552.819906] env[63021]: result = function(*args, **kwargs) [ 552.819906] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 552.819906] env[63021]: return func(*args, **kwargs) [ 552.819906] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 552.819906] env[63021]: raise e [ 552.819906] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 552.819906] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 552.819906] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 552.819906] env[63021]: created_port_ids = self._update_ports_for_instance( [ 552.819906] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 552.819906] env[63021]: with excutils.save_and_reraise_exception(): [ 552.819906] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 552.819906] env[63021]: self.force_reraise() [ 552.819906] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 552.819906] env[63021]: raise self.value [ 552.819906] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 552.819906] env[63021]: updated_port = self._update_port( [ 552.819906] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 552.819906] env[63021]: _ensure_no_port_binding_failure(port) [ 552.819906] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 552.819906] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 552.820633] env[63021]: nova.exception.PortBindingFailed: Binding failed for port 80b3a05d-0a47-4bb2-ae3b-41543c11af6c, please check neutron logs for more information. [ 552.820633] env[63021]: Removing descriptor: 16 [ 552.820633] env[63021]: ERROR nova.compute.manager [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 80b3a05d-0a47-4bb2-ae3b-41543c11af6c, please check neutron logs for more information. [ 552.820633] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Traceback (most recent call last): [ 552.820633] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 552.820633] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] yield resources [ 552.820633] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 552.820633] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] self.driver.spawn(context, instance, image_meta, [ 552.820633] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 552.820633] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] self._vmops.spawn(context, instance, image_meta, injected_files, [ 552.820633] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 552.820633] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] vm_ref = self.build_virtual_machine(instance, [ 552.820941] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 552.820941] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] vif_infos = vmwarevif.get_vif_info(self._session, [ 552.820941] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 552.820941] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] for vif in network_info: [ 552.820941] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 552.820941] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] return self._sync_wrapper(fn, *args, **kwargs) [ 552.820941] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 552.820941] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] self.wait() [ 552.820941] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 552.820941] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] self[:] = self._gt.wait() [ 552.820941] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 552.820941] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] return self._exit_event.wait() [ 552.820941] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 552.821281] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] result = hub.switch() [ 552.821281] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 552.821281] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] return self.greenlet.switch() [ 552.821281] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 552.821281] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] result = function(*args, **kwargs) [ 552.821281] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 552.821281] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] return func(*args, **kwargs) [ 552.821281] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 552.821281] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] raise e [ 552.821281] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 552.821281] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] nwinfo = self.network_api.allocate_for_instance( [ 552.821281] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 552.821281] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] created_port_ids = self._update_ports_for_instance( [ 552.822875] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 552.822875] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] with excutils.save_and_reraise_exception(): [ 552.822875] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 552.822875] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] self.force_reraise() [ 552.822875] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 552.822875] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] raise self.value [ 552.822875] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 552.822875] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] updated_port = self._update_port( [ 552.822875] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 552.822875] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] _ensure_no_port_binding_failure(port) [ 552.822875] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 552.822875] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] raise exception.PortBindingFailed(port_id=port['id']) [ 552.823743] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] nova.exception.PortBindingFailed: Binding failed for port 80b3a05d-0a47-4bb2-ae3b-41543c11af6c, please check neutron logs for more information. [ 552.823743] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] [ 552.823743] env[63021]: INFO nova.compute.manager [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Terminating instance [ 552.824597] env[63021]: DEBUG oslo_concurrency.lockutils [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Acquiring lock "refresh_cache-d629c773-3eb1-4f8d-9888-8bbf7565acea" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 552.834340] env[63021]: DEBUG nova.network.neutron [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Successfully created port: 5adbb74f-e77e-4614-8abf-14b72f2f4658 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 552.838403] env[63021]: DEBUG oslo_concurrency.lockutils [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Releasing lock "refresh_cache-4abfadf8-b4cc-45ef-94b2-fcc6db6afd71" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 552.838403] env[63021]: DEBUG nova.compute.manager [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 552.838640] env[63021]: DEBUG nova.compute.manager [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 552.838857] env[63021]: DEBUG nova.network.neutron [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 552.963056] env[63021]: DEBUG nova.network.neutron [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 552.993988] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Acquiring lock "9d474bd0-e825-49f2-9715-14d85495718e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.993988] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Lock "9d474bd0-e825-49f2-9715-14d85495718e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.114118] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Releasing lock "refresh_cache-d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 553.114568] env[63021]: DEBUG nova.compute.manager [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 553.115214] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 553.115214] env[63021]: DEBUG oslo_concurrency.lockutils [req-e87021f4-bfb6-4c04-a402-85f1240308f9 req-94bbd5d2-36fb-47ae-ab5c-d82adfd3c22c service nova] Acquired lock "refresh_cache-d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 553.115214] env[63021]: DEBUG nova.network.neutron [req-e87021f4-bfb6-4c04-a402-85f1240308f9 req-94bbd5d2-36fb-47ae-ab5c-d82adfd3c22c service nova] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Refreshing network info cache for port a1a43b46-a34b-4ebe-befa-e72aa56aae02 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 553.116287] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2dc64abc-e90a-41d3-af3d-6bae6b675935 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.133733] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c03ee468-d265-47e2-ab79-ebf1309a1adb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.156434] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f could not be found. [ 553.156783] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 553.156983] env[63021]: INFO nova.compute.manager [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 553.157253] env[63021]: DEBUG oslo.service.loopingcall [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 553.157464] env[63021]: DEBUG nova.compute.manager [-] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 553.157558] env[63021]: DEBUG nova.network.neutron [-] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 553.224831] env[63021]: DEBUG nova.network.neutron [-] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 553.237865] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.867s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 553.238599] env[63021]: ERROR nova.compute.manager [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fbbedcd4-9539-41e5-9059-34e6281b2574, please check neutron logs for more information. [ 553.238599] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Traceback (most recent call last): [ 553.238599] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 553.238599] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] self.driver.spawn(context, instance, image_meta, [ 553.238599] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 553.238599] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 553.238599] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 553.238599] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] vm_ref = self.build_virtual_machine(instance, [ 553.238599] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 553.238599] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] vif_infos = vmwarevif.get_vif_info(self._session, [ 553.238599] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 553.238935] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] for vif in network_info: [ 553.238935] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 553.238935] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] return self._sync_wrapper(fn, *args, **kwargs) [ 553.238935] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 553.238935] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] self.wait() [ 553.238935] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 553.238935] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] self[:] = self._gt.wait() [ 553.238935] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 553.238935] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] return self._exit_event.wait() [ 553.238935] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 553.238935] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] result = hub.switch() [ 553.238935] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 553.238935] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] return self.greenlet.switch() [ 553.239288] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 553.239288] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] result = function(*args, **kwargs) [ 553.239288] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 553.239288] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] return func(*args, **kwargs) [ 553.239288] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 553.239288] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] raise e [ 553.239288] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 553.239288] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] nwinfo = self.network_api.allocate_for_instance( [ 553.239288] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 553.239288] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] created_port_ids = self._update_ports_for_instance( [ 553.239288] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 553.239288] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] with excutils.save_and_reraise_exception(): [ 553.239288] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 553.239648] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] self.force_reraise() [ 553.239648] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 553.239648] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] raise self.value [ 553.239648] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 553.239648] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] updated_port = self._update_port( [ 553.239648] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 553.239648] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] _ensure_no_port_binding_failure(port) [ 553.239648] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 553.239648] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] raise exception.PortBindingFailed(port_id=port['id']) [ 553.239648] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] nova.exception.PortBindingFailed: Binding failed for port fbbedcd4-9539-41e5-9059-34e6281b2574, please check neutron logs for more information. [ 553.239648] env[63021]: ERROR nova.compute.manager [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] [ 553.239922] env[63021]: DEBUG nova.compute.utils [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Binding failed for port fbbedcd4-9539-41e5-9059-34e6281b2574, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 553.241446] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 4.615s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.241702] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 553.241702] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63021) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 553.242069] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.454s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.243714] env[63021]: INFO nova.compute.claims [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 553.251103] env[63021]: DEBUG nova.compute.manager [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Build of instance 31b903cb-64b8-46f4-b96f-1d3337ad776d was re-scheduled: Binding failed for port fbbedcd4-9539-41e5-9059-34e6281b2574, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 553.251103] env[63021]: DEBUG nova.compute.manager [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 553.251103] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Acquiring lock "refresh_cache-31b903cb-64b8-46f4-b96f-1d3337ad776d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 553.251103] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Acquired lock "refresh_cache-31b903cb-64b8-46f4-b96f-1d3337ad776d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 553.251448] env[63021]: DEBUG nova.network.neutron [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 553.251448] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fec37ba1-6b15-463b-8edf-a55ae273e216 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.261485] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59e15a95-6c70-4932-8a62-cea93572fe20 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.280829] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a805a80-5351-428e-9383-82fc8c8a8ea3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.288930] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01f5d4d6-2938-4e37-96dd-4d853470b8cf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.325384] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181532MB free_disk=139GB free_vcpus=48 pci_devices=None {{(pid=63021) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 553.325528] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.336592] env[63021]: DEBUG nova.network.neutron [req-9fae82ff-5847-4f9c-a74a-22958fb3767c req-db9a066e-2c9c-4486-913c-3d7d1061f911 service nova] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 553.466911] env[63021]: DEBUG nova.network.neutron [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.645036] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Acquiring lock "ba8e47f7-e10d-439a-9a5a-714655a29dc0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.645333] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Lock "ba8e47f7-e10d-439a-9a5a-714655a29dc0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.669787] env[63021]: DEBUG nova.network.neutron [req-e87021f4-bfb6-4c04-a402-85f1240308f9 req-94bbd5d2-36fb-47ae-ab5c-d82adfd3c22c service nova] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 553.727323] env[63021]: DEBUG nova.network.neutron [-] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.808754] env[63021]: DEBUG nova.network.neutron [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 553.933206] env[63021]: DEBUG nova.network.neutron [req-e87021f4-bfb6-4c04-a402-85f1240308f9 req-94bbd5d2-36fb-47ae-ab5c-d82adfd3c22c service nova] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.951403] env[63021]: DEBUG nova.network.neutron [req-9fae82ff-5847-4f9c-a74a-22958fb3767c req-db9a066e-2c9c-4486-913c-3d7d1061f911 service nova] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.969778] env[63021]: INFO nova.compute.manager [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71] Took 1.13 seconds to deallocate network for instance. [ 554.109911] env[63021]: DEBUG nova.network.neutron [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.229888] env[63021]: INFO nova.compute.manager [-] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Took 1.07 seconds to deallocate network for instance. [ 554.237303] env[63021]: DEBUG nova.compute.claims [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 554.237303] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.437886] env[63021]: DEBUG oslo_concurrency.lockutils [req-e87021f4-bfb6-4c04-a402-85f1240308f9 req-94bbd5d2-36fb-47ae-ab5c-d82adfd3c22c service nova] Releasing lock "refresh_cache-d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 554.457021] env[63021]: DEBUG oslo_concurrency.lockutils [req-9fae82ff-5847-4f9c-a74a-22958fb3767c req-db9a066e-2c9c-4486-913c-3d7d1061f911 service nova] Releasing lock "refresh_cache-d629c773-3eb1-4f8d-9888-8bbf7565acea" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 554.457618] env[63021]: DEBUG oslo_concurrency.lockutils [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Acquired lock "refresh_cache-d629c773-3eb1-4f8d-9888-8bbf7565acea" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 554.458112] env[63021]: DEBUG nova.network.neutron [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 554.519052] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dead5f5f-3e8a-41c7-907f-c1aee5f27027 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.527978] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44555585-f70f-4490-9bc9-dbfd62bc4754 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.560478] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-370c608c-c550-4b56-985f-3e8dbcea3822 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.568227] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81c15ebd-a7ad-46ab-91bf-81708e66526f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.582419] env[63021]: DEBUG nova.compute.provider_tree [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 554.615215] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Releasing lock "refresh_cache-31b903cb-64b8-46f4-b96f-1d3337ad776d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 554.615453] env[63021]: DEBUG nova.compute.manager [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 554.615614] env[63021]: DEBUG nova.compute.manager [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 554.615870] env[63021]: DEBUG nova.network.neutron [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 554.685722] env[63021]: DEBUG nova.network.neutron [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 554.804130] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Acquiring lock "082e6e14-c5f7-4ca4-b739-9d997a77b8f9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.804750] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Lock "082e6e14-c5f7-4ca4-b739-9d997a77b8f9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.018713] env[63021]: INFO nova.scheduler.client.report [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Deleted allocations for instance 4abfadf8-b4cc-45ef-94b2-fcc6db6afd71 [ 555.023636] env[63021]: DEBUG nova.network.neutron [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 555.088661] env[63021]: DEBUG nova.scheduler.client.report [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 555.189890] env[63021]: DEBUG nova.network.neutron [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.466587] env[63021]: DEBUG nova.network.neutron [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.531032] env[63021]: DEBUG oslo_concurrency.lockutils [None req-60d0fc51-8495-4cb0-9f29-d71d53d19812 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Lock "4abfadf8-b4cc-45ef-94b2-fcc6db6afd71" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.336s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 555.595669] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.353s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 555.595829] env[63021]: DEBUG nova.compute.manager [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 555.603024] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.430s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.603024] env[63021]: INFO nova.compute.claims [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 555.696555] env[63021]: INFO nova.compute.manager [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] [instance: 31b903cb-64b8-46f4-b96f-1d3337ad776d] Took 1.08 seconds to deallocate network for instance. [ 555.813644] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Acquiring lock "9f64a5ad-7852-4772-9b8c-2ec1537204f3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.813644] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Lock "9f64a5ad-7852-4772-9b8c-2ec1537204f3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.973595] env[63021]: DEBUG oslo_concurrency.lockutils [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Releasing lock "refresh_cache-d629c773-3eb1-4f8d-9888-8bbf7565acea" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 555.973595] env[63021]: DEBUG nova.compute.manager [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 555.973595] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 555.973595] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b1dea8e0-14ea-4700-b79f-58d3d598d105 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.982840] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1b644f1-b9f0-4be0-8183-30358b3b5b1e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.007825] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d629c773-3eb1-4f8d-9888-8bbf7565acea could not be found. [ 556.008196] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 556.008482] env[63021]: INFO nova.compute.manager [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Took 0.04 seconds to destroy the instance on the hypervisor. [ 556.008833] env[63021]: DEBUG oslo.service.loopingcall [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 556.009168] env[63021]: DEBUG nova.compute.manager [-] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 556.009553] env[63021]: DEBUG nova.network.neutron [-] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 556.034373] env[63021]: DEBUG nova.compute.manager [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 556.063476] env[63021]: DEBUG nova.network.neutron [-] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 556.104684] env[63021]: DEBUG nova.compute.utils [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 556.108520] env[63021]: DEBUG nova.compute.manager [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 556.109507] env[63021]: DEBUG nova.network.neutron [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 556.292559] env[63021]: ERROR nova.compute.manager [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b1544ac9-c61e-4f0a-a311-99a23b19e457, please check neutron logs for more information. [ 556.292559] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 556.292559] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.292559] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 556.292559] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 556.292559] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 556.292559] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 556.292559] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 556.292559] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.292559] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 556.292559] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.292559] env[63021]: ERROR nova.compute.manager raise self.value [ 556.292559] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 556.292559] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 556.292559] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.292559] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 556.293106] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.293106] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 556.293106] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b1544ac9-c61e-4f0a-a311-99a23b19e457, please check neutron logs for more information. [ 556.293106] env[63021]: ERROR nova.compute.manager [ 556.293106] env[63021]: Traceback (most recent call last): [ 556.293106] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 556.293106] env[63021]: listener.cb(fileno) [ 556.293106] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 556.293106] env[63021]: result = function(*args, **kwargs) [ 556.293106] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 556.293106] env[63021]: return func(*args, **kwargs) [ 556.293106] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 556.293106] env[63021]: raise e [ 556.293106] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.293106] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 556.293106] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 556.293106] env[63021]: created_port_ids = self._update_ports_for_instance( [ 556.293106] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 556.293106] env[63021]: with excutils.save_and_reraise_exception(): [ 556.293106] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.293106] env[63021]: self.force_reraise() [ 556.293106] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.293106] env[63021]: raise self.value [ 556.293106] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 556.293106] env[63021]: updated_port = self._update_port( [ 556.293106] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.293106] env[63021]: _ensure_no_port_binding_failure(port) [ 556.293106] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.293106] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 556.293957] env[63021]: nova.exception.PortBindingFailed: Binding failed for port b1544ac9-c61e-4f0a-a311-99a23b19e457, please check neutron logs for more information. [ 556.293957] env[63021]: Removing descriptor: 18 [ 556.293957] env[63021]: ERROR nova.compute.manager [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b1544ac9-c61e-4f0a-a311-99a23b19e457, please check neutron logs for more information. [ 556.293957] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Traceback (most recent call last): [ 556.293957] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 556.293957] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] yield resources [ 556.293957] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 556.293957] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] self.driver.spawn(context, instance, image_meta, [ 556.293957] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 556.293957] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 556.293957] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 556.293957] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] vm_ref = self.build_virtual_machine(instance, [ 556.294278] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 556.294278] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] vif_infos = vmwarevif.get_vif_info(self._session, [ 556.294278] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 556.294278] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] for vif in network_info: [ 556.294278] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 556.294278] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] return self._sync_wrapper(fn, *args, **kwargs) [ 556.294278] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 556.294278] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] self.wait() [ 556.294278] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 556.294278] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] self[:] = self._gt.wait() [ 556.294278] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 556.294278] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] return self._exit_event.wait() [ 556.294278] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 556.294624] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] result = hub.switch() [ 556.294624] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 556.294624] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] return self.greenlet.switch() [ 556.294624] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 556.294624] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] result = function(*args, **kwargs) [ 556.294624] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 556.294624] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] return func(*args, **kwargs) [ 556.294624] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 556.294624] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] raise e [ 556.294624] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.294624] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] nwinfo = self.network_api.allocate_for_instance( [ 556.294624] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 556.294624] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] created_port_ids = self._update_ports_for_instance( [ 556.295218] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 556.295218] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] with excutils.save_and_reraise_exception(): [ 556.295218] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.295218] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] self.force_reraise() [ 556.295218] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.295218] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] raise self.value [ 556.295218] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 556.295218] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] updated_port = self._update_port( [ 556.295218] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.295218] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] _ensure_no_port_binding_failure(port) [ 556.295218] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.295218] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] raise exception.PortBindingFailed(port_id=port['id']) [ 556.295552] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] nova.exception.PortBindingFailed: Binding failed for port b1544ac9-c61e-4f0a-a311-99a23b19e457, please check neutron logs for more information. [ 556.295552] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] [ 556.295552] env[63021]: INFO nova.compute.manager [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Terminating instance [ 556.296944] env[63021]: DEBUG oslo_concurrency.lockutils [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Acquiring lock "refresh_cache-0b443010-4dcf-4a7a-aec3-2d50bd50acc3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 556.297116] env[63021]: DEBUG oslo_concurrency.lockutils [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Acquired lock "refresh_cache-0b443010-4dcf-4a7a-aec3-2d50bd50acc3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.297280] env[63021]: DEBUG nova.network.neutron [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 556.497559] env[63021]: DEBUG nova.policy [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5f8a93bdd7634b7f88823d39054dfa03', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd87b2228ae914fb2bcef9a64d919eb45', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 556.565198] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.565507] env[63021]: DEBUG nova.network.neutron [-] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.609924] env[63021]: DEBUG nova.compute.manager [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 556.746017] env[63021]: INFO nova.scheduler.client.report [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Deleted allocations for instance 31b903cb-64b8-46f4-b96f-1d3337ad776d [ 556.876402] env[63021]: DEBUG nova.network.neutron [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 556.960891] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7273b3ad-3798-48fd-b988-a5e7ac385b31 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.974804] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-034ae3ef-f312-4518-8ce6-1d2680e31a2f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.017358] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-945fe5e5-277c-4934-83fc-520e67421f6a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.028534] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fbe339c-02de-4b6e-86d2-9811ba8867a9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.051892] env[63021]: DEBUG nova.compute.provider_tree [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 557.069242] env[63021]: INFO nova.compute.manager [-] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Took 1.06 seconds to deallocate network for instance. [ 557.072624] env[63021]: DEBUG nova.compute.claims [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 557.072624] env[63021]: DEBUG oslo_concurrency.lockutils [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.257354] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1ead6e3a-4627-4fd5-a9f6-f76fe5ebd03f tempest-TenantUsagesTestJSON-450042440 tempest-TenantUsagesTestJSON-450042440-project-member] Lock "31b903cb-64b8-46f4-b96f-1d3337ad776d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.803s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 557.325788] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Acquiring lock "1fa1c7c7-b9fe-4e60-8306-3c263c517c28" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.327020] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Lock "1fa1c7c7-b9fe-4e60-8306-3c263c517c28" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.376822] env[63021]: DEBUG oslo_concurrency.lockutils [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Acquiring lock "ee9bdcd3-03ac-4093-bf21-877a2a8847a0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.377096] env[63021]: DEBUG oslo_concurrency.lockutils [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Lock "ee9bdcd3-03ac-4093-bf21-877a2a8847a0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.443033] env[63021]: DEBUG nova.network.neutron [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.491202] env[63021]: DEBUG nova.compute.manager [req-c78050a5-fbd0-41b5-85f3-2d21a38835bc req-a64d763e-0829-4548-9c6d-18203024499d service nova] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Received event network-vif-deleted-a1a43b46-a34b-4ebe-befa-e72aa56aae02 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 557.552291] env[63021]: DEBUG nova.scheduler.client.report [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 557.625219] env[63021]: DEBUG nova.compute.manager [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 557.654194] env[63021]: DEBUG nova.compute.manager [req-c7d865cb-1fdc-450f-b076-8553c2781fb9 req-1926fbbf-12d0-4dca-834b-2071c0bbf438 service nova] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Received event network-vif-deleted-80b3a05d-0a47-4bb2-ae3b-41543c11af6c {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 557.665657] env[63021]: DEBUG nova.virt.hardware [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 557.665657] env[63021]: DEBUG nova.virt.hardware [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 557.665899] env[63021]: DEBUG nova.virt.hardware [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 557.666106] env[63021]: DEBUG nova.virt.hardware [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 557.666161] env[63021]: DEBUG nova.virt.hardware [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 557.666351] env[63021]: DEBUG nova.virt.hardware [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 557.666583] env[63021]: DEBUG nova.virt.hardware [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 557.666709] env[63021]: DEBUG nova.virt.hardware [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 557.669036] env[63021]: DEBUG nova.virt.hardware [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 557.669036] env[63021]: DEBUG nova.virt.hardware [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 557.669036] env[63021]: DEBUG nova.virt.hardware [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 557.669036] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec9e6b45-8991-4da3-9177-073e14c926ff {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.683095] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f0dbbcc-c6e7-4926-8eee-cf6452f25d0f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.761784] env[63021]: DEBUG nova.compute.manager [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 557.946196] env[63021]: DEBUG oslo_concurrency.lockutils [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Releasing lock "refresh_cache-0b443010-4dcf-4a7a-aec3-2d50bd50acc3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.946642] env[63021]: DEBUG nova.compute.manager [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 557.949976] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 557.949976] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eb39811f-aa80-459e-954e-53009756d2f9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.959689] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-626f7a92-27f7-4d8f-a4df-1ae97ed00340 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.993669] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0b443010-4dcf-4a7a-aec3-2d50bd50acc3 could not be found. [ 557.994188] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 557.994504] env[63021]: INFO nova.compute.manager [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Took 0.05 seconds to destroy the instance on the hypervisor. [ 557.995526] env[63021]: DEBUG oslo.service.loopingcall [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 557.997016] env[63021]: DEBUG nova.compute.manager [-] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 557.997016] env[63021]: DEBUG nova.network.neutron [-] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 558.035172] env[63021]: DEBUG nova.network.neutron [-] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 558.059722] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.461s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 558.060371] env[63021]: DEBUG nova.compute.manager [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 558.064987] env[63021]: DEBUG oslo_concurrency.lockutils [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.433s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.068819] env[63021]: INFO nova.compute.claims [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 558.291119] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.540529] env[63021]: DEBUG nova.network.neutron [-] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.567099] env[63021]: DEBUG nova.compute.utils [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 558.568854] env[63021]: DEBUG nova.compute.manager [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Not allocating networking since 'none' was specified. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 558.662578] env[63021]: ERROR nova.compute.manager [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5adbb74f-e77e-4614-8abf-14b72f2f4658, please check neutron logs for more information. [ 558.662578] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 558.662578] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 558.662578] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 558.662578] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 558.662578] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 558.662578] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 558.662578] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 558.662578] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 558.662578] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 558.662578] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 558.662578] env[63021]: ERROR nova.compute.manager raise self.value [ 558.662578] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 558.662578] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 558.662578] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 558.662578] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 558.663071] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 558.663071] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 558.663071] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5adbb74f-e77e-4614-8abf-14b72f2f4658, please check neutron logs for more information. [ 558.663071] env[63021]: ERROR nova.compute.manager [ 558.663071] env[63021]: Traceback (most recent call last): [ 558.663071] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 558.663071] env[63021]: listener.cb(fileno) [ 558.663071] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 558.663071] env[63021]: result = function(*args, **kwargs) [ 558.663071] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 558.663071] env[63021]: return func(*args, **kwargs) [ 558.663071] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 558.663071] env[63021]: raise e [ 558.663071] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 558.663071] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 558.663071] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 558.663071] env[63021]: created_port_ids = self._update_ports_for_instance( [ 558.663071] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 558.663071] env[63021]: with excutils.save_and_reraise_exception(): [ 558.663071] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 558.663071] env[63021]: self.force_reraise() [ 558.663071] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 558.663071] env[63021]: raise self.value [ 558.663071] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 558.663071] env[63021]: updated_port = self._update_port( [ 558.663071] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 558.663071] env[63021]: _ensure_no_port_binding_failure(port) [ 558.663071] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 558.663071] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 558.663901] env[63021]: nova.exception.PortBindingFailed: Binding failed for port 5adbb74f-e77e-4614-8abf-14b72f2f4658, please check neutron logs for more information. [ 558.663901] env[63021]: Removing descriptor: 15 [ 558.663901] env[63021]: ERROR nova.compute.manager [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5adbb74f-e77e-4614-8abf-14b72f2f4658, please check neutron logs for more information. [ 558.663901] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Traceback (most recent call last): [ 558.663901] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 558.663901] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] yield resources [ 558.663901] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 558.663901] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] self.driver.spawn(context, instance, image_meta, [ 558.663901] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 558.663901] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] self._vmops.spawn(context, instance, image_meta, injected_files, [ 558.663901] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 558.663901] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] vm_ref = self.build_virtual_machine(instance, [ 558.664244] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 558.664244] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] vif_infos = vmwarevif.get_vif_info(self._session, [ 558.664244] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 558.664244] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] for vif in network_info: [ 558.664244] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 558.664244] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] return self._sync_wrapper(fn, *args, **kwargs) [ 558.664244] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 558.664244] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] self.wait() [ 558.664244] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 558.664244] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] self[:] = self._gt.wait() [ 558.664244] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 558.664244] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] return self._exit_event.wait() [ 558.664244] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 558.664684] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] result = hub.switch() [ 558.664684] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 558.664684] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] return self.greenlet.switch() [ 558.664684] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 558.664684] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] result = function(*args, **kwargs) [ 558.664684] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 558.664684] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] return func(*args, **kwargs) [ 558.664684] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 558.664684] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] raise e [ 558.664684] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 558.664684] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] nwinfo = self.network_api.allocate_for_instance( [ 558.664684] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 558.664684] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] created_port_ids = self._update_ports_for_instance( [ 558.665043] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 558.665043] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] with excutils.save_and_reraise_exception(): [ 558.665043] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 558.665043] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] self.force_reraise() [ 558.665043] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 558.665043] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] raise self.value [ 558.665043] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 558.665043] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] updated_port = self._update_port( [ 558.665043] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 558.665043] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] _ensure_no_port_binding_failure(port) [ 558.665043] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 558.665043] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] raise exception.PortBindingFailed(port_id=port['id']) [ 558.665432] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] nova.exception.PortBindingFailed: Binding failed for port 5adbb74f-e77e-4614-8abf-14b72f2f4658, please check neutron logs for more information. [ 558.665432] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] [ 558.665432] env[63021]: INFO nova.compute.manager [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Terminating instance [ 558.669645] env[63021]: DEBUG oslo_concurrency.lockutils [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Acquiring lock "refresh_cache-7d98e2e4-33bc-4e6b-af6f-31c18f34b9af" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 558.669645] env[63021]: DEBUG oslo_concurrency.lockutils [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Acquired lock "refresh_cache-7d98e2e4-33bc-4e6b-af6f-31c18f34b9af" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 558.669645] env[63021]: DEBUG nova.network.neutron [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 558.704987] env[63021]: DEBUG nova.network.neutron [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Successfully created port: f2cc2dc3-c7ef-4dcc-bbf3-c9154df16cc6 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 559.045180] env[63021]: INFO nova.compute.manager [-] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Took 1.05 seconds to deallocate network for instance. [ 559.048079] env[63021]: DEBUG nova.compute.claims [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 559.049950] env[63021]: DEBUG oslo_concurrency.lockutils [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.071801] env[63021]: DEBUG nova.compute.manager [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 559.220166] env[63021]: DEBUG nova.network.neutron [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 559.358890] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7dffef4-ea30-48d5-ad79-44a5c7f89aca {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.370027] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-531dd132-80ba-46c5-b4ed-c4cebc28ccbe {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.407419] env[63021]: DEBUG nova.network.neutron [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.409514] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2902e23-8b61-4669-87be-7748e692e4e8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.418425] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-804b1ffe-85a0-4d7c-b2d6-e5961d8a4361 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.438150] env[63021]: DEBUG nova.compute.provider_tree [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 559.913374] env[63021]: DEBUG oslo_concurrency.lockutils [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Releasing lock "refresh_cache-7d98e2e4-33bc-4e6b-af6f-31c18f34b9af" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.914053] env[63021]: DEBUG nova.compute.manager [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 559.914246] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 559.914417] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1bd1082f-d8de-412b-b3b3-752000e3b8f2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.927567] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d333aa5-4743-43f2-af31-9fc11836573f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.940891] env[63021]: DEBUG nova.scheduler.client.report [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 559.955942] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af could not be found. [ 559.956382] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 559.956382] env[63021]: INFO nova.compute.manager [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Took 0.04 seconds to destroy the instance on the hypervisor. [ 559.956573] env[63021]: DEBUG oslo.service.loopingcall [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 559.956632] env[63021]: DEBUG nova.compute.manager [-] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 559.956719] env[63021]: DEBUG nova.network.neutron [-] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 560.024029] env[63021]: DEBUG nova.network.neutron [-] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 560.089617] env[63021]: DEBUG nova.compute.manager [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 560.123640] env[63021]: DEBUG nova.virt.hardware [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 560.123928] env[63021]: DEBUG nova.virt.hardware [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 560.124036] env[63021]: DEBUG nova.virt.hardware [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 560.124225] env[63021]: DEBUG nova.virt.hardware [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 560.124367] env[63021]: DEBUG nova.virt.hardware [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 560.124514] env[63021]: DEBUG nova.virt.hardware [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 560.124728] env[63021]: DEBUG nova.virt.hardware [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 560.124881] env[63021]: DEBUG nova.virt.hardware [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 560.125052] env[63021]: DEBUG nova.virt.hardware [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 560.125211] env[63021]: DEBUG nova.virt.hardware [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 560.125374] env[63021]: DEBUG nova.virt.hardware [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 560.126257] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81f1b2d1-47d5-4569-9b8b-4880d6f5f3ab {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.137917] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7d66de8-8c76-4d5b-8c3c-494a5da0d7e1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.156048] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Instance VIF info [] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 560.164669] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 560.165048] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f3e4b3a5-eddd-41dc-a356-9b5ba2d0f7db {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.179557] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Created folder: OpenStack in parent group-v4. [ 560.179755] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Creating folder: Project (5ca38529140c47cc818e90b15606fcda). Parent ref: group-v277447. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 560.179999] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-428b37a2-b0f6-4cef-bb7c-11daf492922a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.191665] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Created folder: Project (5ca38529140c47cc818e90b15606fcda) in parent group-v277447. [ 560.191665] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Creating folder: Instances. Parent ref: group-v277448. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 560.191665] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e5505819-29a4-4779-bb9a-61bbb081a85d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.201543] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Created folder: Instances in parent group-v277448. [ 560.202194] env[63021]: DEBUG oslo.service.loopingcall [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 560.202194] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 560.202304] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ed0f1736-5d4c-4545-8591-a99dcb206570 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.221947] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 560.221947] env[63021]: value = "task-1293522" [ 560.221947] env[63021]: _type = "Task" [ 560.221947] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 560.230332] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293522, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 560.447747] env[63021]: DEBUG oslo_concurrency.lockutils [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.382s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.448675] env[63021]: DEBUG nova.compute.manager [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 560.454569] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.685s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.454569] env[63021]: INFO nova.compute.claims [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 560.526802] env[63021]: DEBUG nova.network.neutron [-] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.734449] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293522, 'name': CreateVM_Task, 'duration_secs': 0.363621} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 560.735203] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 560.735780] env[63021]: DEBUG oslo_vmware.service [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3224c59-6c7b-4b66-9e27-347668cde152 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.742975] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.743035] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.745536] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 560.745826] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0021378-ece1-40a6-a3f6-743d9c5d36b3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.751033] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 560.751033] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52a3ea3c-3a4e-af84-d88e-adc3618104dd" [ 560.751033] env[63021]: _type = "Task" [ 560.751033] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 560.764017] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52a3ea3c-3a4e-af84-d88e-adc3618104dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 560.799126] env[63021]: DEBUG nova.compute.manager [req-d907e531-4add-4dbe-aba2-ba3cffc94d25 req-07ead579-134f-4c20-95dd-999d72f55751 service nova] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Received event network-changed-b1544ac9-c61e-4f0a-a311-99a23b19e457 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 560.799444] env[63021]: DEBUG nova.compute.manager [req-d907e531-4add-4dbe-aba2-ba3cffc94d25 req-07ead579-134f-4c20-95dd-999d72f55751 service nova] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Refreshing instance network info cache due to event network-changed-b1544ac9-c61e-4f0a-a311-99a23b19e457. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 560.799799] env[63021]: DEBUG oslo_concurrency.lockutils [req-d907e531-4add-4dbe-aba2-ba3cffc94d25 req-07ead579-134f-4c20-95dd-999d72f55751 service nova] Acquiring lock "refresh_cache-0b443010-4dcf-4a7a-aec3-2d50bd50acc3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.799907] env[63021]: DEBUG oslo_concurrency.lockutils [req-d907e531-4add-4dbe-aba2-ba3cffc94d25 req-07ead579-134f-4c20-95dd-999d72f55751 service nova] Acquired lock "refresh_cache-0b443010-4dcf-4a7a-aec3-2d50bd50acc3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.800158] env[63021]: DEBUG nova.network.neutron [req-d907e531-4add-4dbe-aba2-ba3cffc94d25 req-07ead579-134f-4c20-95dd-999d72f55751 service nova] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Refreshing network info cache for port b1544ac9-c61e-4f0a-a311-99a23b19e457 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 560.960653] env[63021]: DEBUG nova.compute.utils [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 560.970831] env[63021]: DEBUG nova.compute.manager [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 560.971188] env[63021]: DEBUG nova.network.neutron [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 561.034234] env[63021]: INFO nova.compute.manager [-] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Took 1.08 seconds to deallocate network for instance. [ 561.035259] env[63021]: DEBUG nova.compute.claims [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 561.035259] env[63021]: DEBUG oslo_concurrency.lockutils [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.222218] env[63021]: DEBUG nova.policy [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '50d5450d7d1541388321de2146612925', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '44ac682161f24e29a7592fac57e1443b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 561.265254] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 561.265449] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 561.265746] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 561.265892] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 561.266382] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 561.266639] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d9741a36-825d-403d-842a-888f332bde71 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.283992] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 561.284232] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 561.285315] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e827f26c-232f-4b32-90b3-3cfec7556e30 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.293106] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37034d13-331a-4304-a02e-ba6cc3be4990 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.303251] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 561.303251] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5212c513-de26-95a3-65a2-4bd772cbfac4" [ 561.303251] env[63021]: _type = "Task" [ 561.303251] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 561.316837] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5212c513-de26-95a3-65a2-4bd772cbfac4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 561.371201] env[63021]: DEBUG nova.network.neutron [req-d907e531-4add-4dbe-aba2-ba3cffc94d25 req-07ead579-134f-4c20-95dd-999d72f55751 service nova] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 561.466285] env[63021]: DEBUG nova.compute.manager [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 561.746979] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff7416b1-3803-4e84-8e2a-43a1cb1b5900 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.757408] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9740955-6a05-4485-9236-146b4c75eaaa {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.797688] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92325af5-db26-40a9-b931-63e4c7356d35 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.813010] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f63bf12-926e-4a37-a8cd-1664d8009968 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.833941] env[63021]: DEBUG nova.compute.provider_tree [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 561.838612] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Preparing fetch location {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 561.838878] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Creating directory with path [datastore1] vmware_temp/53b0b3b3-ed73-42e5-8ca1-e9ffee0dd667/e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 561.839139] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-111a0b2d-ea16-4e2d-81f7-3d92d3643894 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.865373] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Created directory with path [datastore1] vmware_temp/53b0b3b3-ed73-42e5-8ca1-e9ffee0dd667/e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 561.865373] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Fetch image to [datastore1] vmware_temp/53b0b3b3-ed73-42e5-8ca1-e9ffee0dd667/e624225d-af0f-415c-90cc-d4b228c479f6/tmp-sparse.vmdk {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 561.865373] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Downloading image file data e624225d-af0f-415c-90cc-d4b228c479f6 to [datastore1] vmware_temp/53b0b3b3-ed73-42e5-8ca1-e9ffee0dd667/e624225d-af0f-415c-90cc-d4b228c479f6/tmp-sparse.vmdk on the data store datastore1 {{(pid=63021) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 561.865590] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de295c7-957f-4bd8-bd74-37592ce6c11b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.877421] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45e24fe4-e812-47e0-bcb1-8566d1811e91 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.887615] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9791615e-9c1a-4b48-bdcb-4f32c196c1b9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.925320] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f7ac741-0d84-4f88-a1f2-de830aeea944 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.932624] env[63021]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-8494c047-8b26-41a1-9170-b53ca8aab354 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.949065] env[63021]: DEBUG nova.network.neutron [req-d907e531-4add-4dbe-aba2-ba3cffc94d25 req-07ead579-134f-4c20-95dd-999d72f55751 service nova] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.031029] env[63021]: DEBUG nova.virt.vmwareapi.images [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Downloading image file data e624225d-af0f-415c-90cc-d4b228c479f6 to the data store datastore1 {{(pid=63021) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 562.148763] env[63021]: DEBUG oslo_vmware.rw_handles [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/53b0b3b3-ed73-42e5-8ca1-e9ffee0dd667/e624225d-af0f-415c-90cc-d4b228c479f6/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=63021) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 562.341439] env[63021]: DEBUG nova.scheduler.client.report [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 562.452179] env[63021]: DEBUG oslo_concurrency.lockutils [req-d907e531-4add-4dbe-aba2-ba3cffc94d25 req-07ead579-134f-4c20-95dd-999d72f55751 service nova] Releasing lock "refresh_cache-0b443010-4dcf-4a7a-aec3-2d50bd50acc3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 562.452316] env[63021]: DEBUG nova.compute.manager [req-d907e531-4add-4dbe-aba2-ba3cffc94d25 req-07ead579-134f-4c20-95dd-999d72f55751 service nova] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Received event network-vif-deleted-b1544ac9-c61e-4f0a-a311-99a23b19e457 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 562.452503] env[63021]: DEBUG nova.compute.manager [req-d907e531-4add-4dbe-aba2-ba3cffc94d25 req-07ead579-134f-4c20-95dd-999d72f55751 service nova] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Received event network-changed-5adbb74f-e77e-4614-8abf-14b72f2f4658 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 562.452683] env[63021]: DEBUG nova.compute.manager [req-d907e531-4add-4dbe-aba2-ba3cffc94d25 req-07ead579-134f-4c20-95dd-999d72f55751 service nova] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Refreshing instance network info cache due to event network-changed-5adbb74f-e77e-4614-8abf-14b72f2f4658. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 562.452930] env[63021]: DEBUG oslo_concurrency.lockutils [req-d907e531-4add-4dbe-aba2-ba3cffc94d25 req-07ead579-134f-4c20-95dd-999d72f55751 service nova] Acquiring lock "refresh_cache-7d98e2e4-33bc-4e6b-af6f-31c18f34b9af" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 562.453097] env[63021]: DEBUG oslo_concurrency.lockutils [req-d907e531-4add-4dbe-aba2-ba3cffc94d25 req-07ead579-134f-4c20-95dd-999d72f55751 service nova] Acquired lock "refresh_cache-7d98e2e4-33bc-4e6b-af6f-31c18f34b9af" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 562.453257] env[63021]: DEBUG nova.network.neutron [req-d907e531-4add-4dbe-aba2-ba3cffc94d25 req-07ead579-134f-4c20-95dd-999d72f55751 service nova] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Refreshing network info cache for port 5adbb74f-e77e-4614-8abf-14b72f2f4658 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 562.486267] env[63021]: DEBUG nova.compute.manager [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 562.530600] env[63021]: DEBUG nova.virt.hardware [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 562.531114] env[63021]: DEBUG nova.virt.hardware [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 562.531349] env[63021]: DEBUG nova.virt.hardware [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 562.533395] env[63021]: DEBUG nova.virt.hardware [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 562.533395] env[63021]: DEBUG nova.virt.hardware [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 562.533395] env[63021]: DEBUG nova.virt.hardware [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 562.533395] env[63021]: DEBUG nova.virt.hardware [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 562.533395] env[63021]: DEBUG nova.virt.hardware [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 562.533747] env[63021]: DEBUG nova.virt.hardware [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 562.533747] env[63021]: DEBUG nova.virt.hardware [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 562.533747] env[63021]: DEBUG nova.virt.hardware [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 562.537823] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-651241f3-7c91-4b6c-bcaa-f2b448bc8129 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.553017] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f78ee706-908b-4c4b-b1a7-3c7cd6841bd4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.825168] env[63021]: DEBUG oslo_vmware.rw_handles [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Completed reading data from the image iterator. {{(pid=63021) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 562.825450] env[63021]: DEBUG oslo_vmware.rw_handles [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Closing write handle for https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/53b0b3b3-ed73-42e5-8ca1-e9ffee0dd667/e624225d-af0f-415c-90cc-d4b228c479f6/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=63021) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 562.841432] env[63021]: DEBUG oslo_concurrency.lockutils [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Acquiring lock "e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.841609] env[63021]: DEBUG oslo_concurrency.lockutils [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Lock "e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.845974] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.393s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 562.846449] env[63021]: DEBUG nova.compute.manager [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 562.849912] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 9.524s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.949027] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "17ff6693-c323-478b-8b61-776c3b41f77e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.950977] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "17ff6693-c323-478b-8b61-776c3b41f77e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.026770] env[63021]: DEBUG nova.network.neutron [req-d907e531-4add-4dbe-aba2-ba3cffc94d25 req-07ead579-134f-4c20-95dd-999d72f55751 service nova] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 563.114817] env[63021]: DEBUG nova.virt.vmwareapi.images [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Downloaded image file data e624225d-af0f-415c-90cc-d4b228c479f6 to vmware_temp/53b0b3b3-ed73-42e5-8ca1-e9ffee0dd667/e624225d-af0f-415c-90cc-d4b228c479f6/tmp-sparse.vmdk on the data store datastore1 {{(pid=63021) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 563.116628] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Caching image {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 563.116884] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Copying Virtual Disk [datastore1] vmware_temp/53b0b3b3-ed73-42e5-8ca1-e9ffee0dd667/e624225d-af0f-415c-90cc-d4b228c479f6/tmp-sparse.vmdk to [datastore1] vmware_temp/53b0b3b3-ed73-42e5-8ca1-e9ffee0dd667/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 563.118231] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1ab62ca8-4afe-4896-a584-2daa26c91012 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.126225] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 563.126225] env[63021]: value = "task-1293523" [ 563.126225] env[63021]: _type = "Task" [ 563.126225] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 563.140028] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293523, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 563.255980] env[63021]: DEBUG oslo_concurrency.lockutils [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Acquiring lock "efd0e375-3134-4984-8f82-48b53c65e26c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.256185] env[63021]: DEBUG oslo_concurrency.lockutils [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Lock "efd0e375-3134-4984-8f82-48b53c65e26c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.351269] env[63021]: DEBUG nova.compute.utils [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 563.352738] env[63021]: DEBUG nova.compute.manager [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Not allocating networking since 'none' was specified. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 563.498103] env[63021]: DEBUG nova.network.neutron [req-d907e531-4add-4dbe-aba2-ba3cffc94d25 req-07ead579-134f-4c20-95dd-999d72f55751 service nova] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.507064] env[63021]: ERROR nova.compute.manager [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f2cc2dc3-c7ef-4dcc-bbf3-c9154df16cc6, please check neutron logs for more information. [ 563.507064] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 563.507064] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 563.507064] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 563.507064] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 563.507064] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 563.507064] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 563.507064] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 563.507064] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.507064] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 563.507064] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.507064] env[63021]: ERROR nova.compute.manager raise self.value [ 563.507064] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 563.507064] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 563.507064] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.507064] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 563.507508] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.507508] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 563.507508] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f2cc2dc3-c7ef-4dcc-bbf3-c9154df16cc6, please check neutron logs for more information. [ 563.507508] env[63021]: ERROR nova.compute.manager [ 563.507508] env[63021]: Traceback (most recent call last): [ 563.507508] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 563.507508] env[63021]: listener.cb(fileno) [ 563.507508] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 563.507508] env[63021]: result = function(*args, **kwargs) [ 563.507508] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 563.507508] env[63021]: return func(*args, **kwargs) [ 563.507508] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 563.507508] env[63021]: raise e [ 563.507508] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 563.507508] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 563.507508] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 563.507508] env[63021]: created_port_ids = self._update_ports_for_instance( [ 563.507508] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 563.507508] env[63021]: with excutils.save_and_reraise_exception(): [ 563.507508] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.507508] env[63021]: self.force_reraise() [ 563.507508] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.507508] env[63021]: raise self.value [ 563.507508] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 563.507508] env[63021]: updated_port = self._update_port( [ 563.507508] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.507508] env[63021]: _ensure_no_port_binding_failure(port) [ 563.507508] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.507508] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 563.508423] env[63021]: nova.exception.PortBindingFailed: Binding failed for port f2cc2dc3-c7ef-4dcc-bbf3-c9154df16cc6, please check neutron logs for more information. [ 563.508423] env[63021]: Removing descriptor: 17 [ 563.508423] env[63021]: ERROR nova.compute.manager [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f2cc2dc3-c7ef-4dcc-bbf3-c9154df16cc6, please check neutron logs for more information. [ 563.508423] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Traceback (most recent call last): [ 563.508423] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 563.508423] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] yield resources [ 563.508423] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 563.508423] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] self.driver.spawn(context, instance, image_meta, [ 563.508423] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 563.508423] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] self._vmops.spawn(context, instance, image_meta, injected_files, [ 563.508423] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 563.508423] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] vm_ref = self.build_virtual_machine(instance, [ 563.508741] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 563.508741] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] vif_infos = vmwarevif.get_vif_info(self._session, [ 563.508741] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 563.508741] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] for vif in network_info: [ 563.508741] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 563.508741] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] return self._sync_wrapper(fn, *args, **kwargs) [ 563.508741] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 563.508741] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] self.wait() [ 563.508741] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 563.508741] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] self[:] = self._gt.wait() [ 563.508741] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 563.508741] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] return self._exit_event.wait() [ 563.508741] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 563.509076] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] result = hub.switch() [ 563.509076] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 563.509076] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] return self.greenlet.switch() [ 563.509076] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 563.509076] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] result = function(*args, **kwargs) [ 563.509076] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 563.509076] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] return func(*args, **kwargs) [ 563.509076] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 563.509076] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] raise e [ 563.509076] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 563.509076] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] nwinfo = self.network_api.allocate_for_instance( [ 563.509076] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 563.509076] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] created_port_ids = self._update_ports_for_instance( [ 563.509515] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 563.509515] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] with excutils.save_and_reraise_exception(): [ 563.509515] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.509515] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] self.force_reraise() [ 563.509515] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.509515] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] raise self.value [ 563.509515] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 563.509515] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] updated_port = self._update_port( [ 563.509515] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.509515] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] _ensure_no_port_binding_failure(port) [ 563.509515] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.509515] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] raise exception.PortBindingFailed(port_id=port['id']) [ 563.509821] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] nova.exception.PortBindingFailed: Binding failed for port f2cc2dc3-c7ef-4dcc-bbf3-c9154df16cc6, please check neutron logs for more information. [ 563.509821] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] [ 563.509821] env[63021]: INFO nova.compute.manager [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Terminating instance [ 563.513814] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Acquiring lock "refresh_cache-ea199990-9638-4e94-87fe-bd4b7f020595" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 563.513982] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Acquired lock "refresh_cache-ea199990-9638-4e94-87fe-bd4b7f020595" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 563.514165] env[63021]: DEBUG nova.network.neutron [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 563.637944] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293523, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 563.733522] env[63021]: DEBUG nova.network.neutron [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Successfully created port: 60d66191-46ca-4e31-8dda-90f2a0c92b59 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 563.865784] env[63021]: DEBUG nova.compute.manager [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 563.895096] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 563.895254] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance d629c773-3eb1-4f8d-9888-8bbf7565acea actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 563.895387] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 0b443010-4dcf-4a7a-aec3-2d50bd50acc3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 563.895503] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 563.895613] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance ea199990-9638-4e94-87fe-bd4b7f020595 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 563.895733] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 0950cbd3-0708-4994-8d2d-053aecc06cdf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 563.895852] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 563.896785] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 564.001532] env[63021]: DEBUG oslo_concurrency.lockutils [req-d907e531-4add-4dbe-aba2-ba3cffc94d25 req-07ead579-134f-4c20-95dd-999d72f55751 service nova] Releasing lock "refresh_cache-7d98e2e4-33bc-4e6b-af6f-31c18f34b9af" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 564.001868] env[63021]: DEBUG nova.compute.manager [req-d907e531-4add-4dbe-aba2-ba3cffc94d25 req-07ead579-134f-4c20-95dd-999d72f55751 service nova] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Received event network-vif-deleted-5adbb74f-e77e-4614-8abf-14b72f2f4658 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 564.058251] env[63021]: DEBUG nova.network.neutron [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 564.143609] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293523, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.987163} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 564.143609] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Copied Virtual Disk [datastore1] vmware_temp/53b0b3b3-ed73-42e5-8ca1-e9ffee0dd667/e624225d-af0f-415c-90cc-d4b228c479f6/tmp-sparse.vmdk to [datastore1] vmware_temp/53b0b3b3-ed73-42e5-8ca1-e9ffee0dd667/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 564.143609] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Deleting the datastore file [datastore1] vmware_temp/53b0b3b3-ed73-42e5-8ca1-e9ffee0dd667/e624225d-af0f-415c-90cc-d4b228c479f6/tmp-sparse.vmdk {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 564.144107] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b27bef0e-7342-4d9c-8b0c-4e6669d27e72 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.151375] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 564.151375] env[63021]: value = "task-1293524" [ 564.151375] env[63021]: _type = "Task" [ 564.151375] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 564.164408] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293524, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 564.285204] env[63021]: DEBUG nova.network.neutron [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.399912] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 9d474bd0-e825-49f2-9715-14d85495718e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 564.667643] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293524, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 564.789164] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Releasing lock "refresh_cache-ea199990-9638-4e94-87fe-bd4b7f020595" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 564.790211] env[63021]: DEBUG nova.compute.manager [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 564.790211] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 564.790211] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-be44b50a-7e0f-4357-8fa2-1d8068bf195f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.801441] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84e9de56-8d86-4305-ac18-2519ecc7cc27 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.826280] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ea199990-9638-4e94-87fe-bd4b7f020595 could not be found. [ 564.826531] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 564.826691] env[63021]: INFO nova.compute.manager [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Took 0.04 seconds to destroy the instance on the hypervisor. [ 564.826945] env[63021]: DEBUG oslo.service.loopingcall [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 564.827188] env[63021]: DEBUG nova.compute.manager [-] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 564.827294] env[63021]: DEBUG nova.network.neutron [-] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 564.877755] env[63021]: DEBUG nova.compute.manager [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 564.906016] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance ba8e47f7-e10d-439a-9a5a-714655a29dc0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 564.916332] env[63021]: DEBUG nova.virt.hardware [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 564.916614] env[63021]: DEBUG nova.virt.hardware [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 564.916745] env[63021]: DEBUG nova.virt.hardware [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 564.916924] env[63021]: DEBUG nova.virt.hardware [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 564.918680] env[63021]: DEBUG nova.virt.hardware [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 564.918963] env[63021]: DEBUG nova.virt.hardware [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 564.919212] env[63021]: DEBUG nova.virt.hardware [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 564.919380] env[63021]: DEBUG nova.virt.hardware [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 564.919563] env[63021]: DEBUG nova.virt.hardware [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 564.919730] env[63021]: DEBUG nova.virt.hardware [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 564.919918] env[63021]: DEBUG nova.virt.hardware [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 564.920983] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a659210-ab84-40b5-b5b9-a2164730574b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.931867] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e5e0270-80d1-4c5a-878b-bd8332fb064c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.937362] env[63021]: DEBUG nova.network.neutron [-] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 564.954991] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Instance VIF info [] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 564.962192] env[63021]: DEBUG oslo.service.loopingcall [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 564.962924] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 564.963196] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7de8fb64-73dc-426e-9926-5d5ab410c6e2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.984348] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 564.984348] env[63021]: value = "task-1293525" [ 564.984348] env[63021]: _type = "Task" [ 564.984348] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 564.994280] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293525, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 565.167497] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293524, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.67467} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 565.167497] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 565.167497] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Moving file from [datastore1] vmware_temp/53b0b3b3-ed73-42e5-8ca1-e9ffee0dd667/e624225d-af0f-415c-90cc-d4b228c479f6 to [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6. {{(pid=63021) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 565.168205] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-f9a2bf0b-cf95-4ef2-a107-02a3280c5a43 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.176452] env[63021]: DEBUG nova.compute.manager [req-bed31e40-8efc-42ee-b28f-d2e8f99be1df req-acfa62a0-7f4b-4410-aaae-f4af35a5d557 service nova] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Received event network-changed-f2cc2dc3-c7ef-4dcc-bbf3-c9154df16cc6 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 565.176452] env[63021]: DEBUG nova.compute.manager [req-bed31e40-8efc-42ee-b28f-d2e8f99be1df req-acfa62a0-7f4b-4410-aaae-f4af35a5d557 service nova] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Refreshing instance network info cache due to event network-changed-f2cc2dc3-c7ef-4dcc-bbf3-c9154df16cc6. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 565.176452] env[63021]: DEBUG oslo_concurrency.lockutils [req-bed31e40-8efc-42ee-b28f-d2e8f99be1df req-acfa62a0-7f4b-4410-aaae-f4af35a5d557 service nova] Acquiring lock "refresh_cache-ea199990-9638-4e94-87fe-bd4b7f020595" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 565.176452] env[63021]: DEBUG oslo_concurrency.lockutils [req-bed31e40-8efc-42ee-b28f-d2e8f99be1df req-acfa62a0-7f4b-4410-aaae-f4af35a5d557 service nova] Acquired lock "refresh_cache-ea199990-9638-4e94-87fe-bd4b7f020595" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 565.176452] env[63021]: DEBUG nova.network.neutron [req-bed31e40-8efc-42ee-b28f-d2e8f99be1df req-acfa62a0-7f4b-4410-aaae-f4af35a5d557 service nova] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Refreshing network info cache for port f2cc2dc3-c7ef-4dcc-bbf3-c9154df16cc6 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 565.185439] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 565.185439] env[63021]: value = "task-1293526" [ 565.185439] env[63021]: _type = "Task" [ 565.185439] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 565.202611] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293526, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 565.352884] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Acquiring lock "92280cdf-bad2-4893-9a37-706cb1bcbac4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.353355] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Lock "92280cdf-bad2-4893-9a37-706cb1bcbac4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.411398] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 082e6e14-c5f7-4ca4-b739-9d997a77b8f9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 565.440996] env[63021]: DEBUG nova.network.neutron [-] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.495740] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293525, 'name': CreateVM_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 565.699748] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293526, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.029048} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 565.700036] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] File moved {{(pid=63021) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 565.700358] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Cleaning up location [datastore1] vmware_temp/53b0b3b3-ed73-42e5-8ca1-e9ffee0dd667 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 565.700423] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Deleting the datastore file [datastore1] vmware_temp/53b0b3b3-ed73-42e5-8ca1-e9ffee0dd667 {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 565.700706] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c26b2d4e-6059-4957-bec8-955d9bb7a7db {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.710950] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 565.710950] env[63021]: value = "task-1293527" [ 565.710950] env[63021]: _type = "Task" [ 565.710950] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 565.719504] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293527, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 565.744112] env[63021]: DEBUG nova.network.neutron [req-bed31e40-8efc-42ee-b28f-d2e8f99be1df req-acfa62a0-7f4b-4410-aaae-f4af35a5d557 service nova] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 565.774576] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Acquiring lock "cee88fa5-c98e-405f-b41c-62d6661b983c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.774576] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Lock "cee88fa5-c98e-405f-b41c-62d6661b983c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.920152] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 9f64a5ad-7852-4772-9b8c-2ec1537204f3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 565.942696] env[63021]: INFO nova.compute.manager [-] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Took 1.12 seconds to deallocate network for instance. [ 565.945909] env[63021]: DEBUG nova.compute.claims [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 565.946096] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.996178] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293525, 'name': CreateVM_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.142634] env[63021]: DEBUG nova.network.neutron [req-bed31e40-8efc-42ee-b28f-d2e8f99be1df req-acfa62a0-7f4b-4410-aaae-f4af35a5d557 service nova] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.223380] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293527, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.038577} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 566.224017] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 566.225229] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba1e9168-1c6d-4caa-be46-b5917bd1ace8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.233931] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 566.233931] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]524198be-f215-2bbf-0fe7-9c4ceea0fe97" [ 566.233931] env[63021]: _type = "Task" [ 566.233931] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 566.249528] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]524198be-f215-2bbf-0fe7-9c4ceea0fe97, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.427624] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 1fa1c7c7-b9fe-4e60-8306-3c263c517c28 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 566.497248] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293525, 'name': CreateVM_Task, 'duration_secs': 1.444603} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 566.497419] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 566.499745] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.499745] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.499745] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 566.499745] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67cfd624-31ed-483b-92d4-e4f1300e54c1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.504030] env[63021]: DEBUG oslo_vmware.api [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 566.504030] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52bd5bb3-cfde-b6ef-e28e-7e0f901e82da" [ 566.504030] env[63021]: _type = "Task" [ 566.504030] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 566.515483] env[63021]: DEBUG oslo_vmware.api [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52bd5bb3-cfde-b6ef-e28e-7e0f901e82da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.645351] env[63021]: DEBUG oslo_concurrency.lockutils [req-bed31e40-8efc-42ee-b28f-d2e8f99be1df req-acfa62a0-7f4b-4410-aaae-f4af35a5d557 service nova] Releasing lock "refresh_cache-ea199990-9638-4e94-87fe-bd4b7f020595" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 566.748522] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]524198be-f215-2bbf-0fe7-9c4ceea0fe97, 'name': SearchDatastore_Task, 'duration_secs': 0.035834} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 566.748837] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 566.749190] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 0950cbd3-0708-4994-8d2d-053aecc06cdf/0950cbd3-0708-4994-8d2d-053aecc06cdf.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 566.749762] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-deb00b20-9490-4ac0-a5ad-6fb9ecb92af7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.756516] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 566.756516] env[63021]: value = "task-1293528" [ 566.756516] env[63021]: _type = "Task" [ 566.756516] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 566.766466] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293528, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.931921] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance ee9bdcd3-03ac-4093-bf21-877a2a8847a0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 567.019567] env[63021]: DEBUG oslo_vmware.api [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52bd5bb3-cfde-b6ef-e28e-7e0f901e82da, 'name': SearchDatastore_Task, 'duration_secs': 0.013037} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 567.019567] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 567.019705] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 567.019928] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 567.020521] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 567.020521] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 567.020521] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e03693dd-090e-4cbb-91ea-f5077c8e85c4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.084926] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 567.085027] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 567.085763] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bfa3d896-bcd8-4a0d-a967-caf960a99c13 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.094800] env[63021]: DEBUG oslo_vmware.api [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 567.094800] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52ade685-e132-ec29-4654-973ac32c592a" [ 567.094800] env[63021]: _type = "Task" [ 567.094800] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 567.106521] env[63021]: DEBUG oslo_vmware.api [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52ade685-e132-ec29-4654-973ac32c592a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 567.275430] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293528, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 567.437291] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 567.596138] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Acquiring lock "ab11dead-3fe6-46d3-92e1-cdca84ebe508" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.596433] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Lock "ab11dead-3fe6-46d3-92e1-cdca84ebe508" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.611160] env[63021]: DEBUG oslo_vmware.api [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52ade685-e132-ec29-4654-973ac32c592a, 'name': SearchDatastore_Task, 'duration_secs': 0.009506} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 567.611160] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4864604-d1a1-4fb3-b086-5179b0255a51 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.618204] env[63021]: DEBUG oslo_vmware.api [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 567.618204] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5242bd31-0da3-7ca4-4578-287e7bc36236" [ 567.618204] env[63021]: _type = "Task" [ 567.618204] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 567.632031] env[63021]: DEBUG oslo_vmware.api [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5242bd31-0da3-7ca4-4578-287e7bc36236, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 567.772519] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293528, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.903817} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 567.773577] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 0950cbd3-0708-4994-8d2d-053aecc06cdf/0950cbd3-0708-4994-8d2d-053aecc06cdf.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 567.773805] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 567.774061] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bea244a8-b9b9-4362-af81-22fe7554c5eb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.783831] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 567.783831] env[63021]: value = "task-1293529" [ 567.783831] env[63021]: _type = "Task" [ 567.783831] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 567.797530] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293529, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 567.940361] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 17ff6693-c323-478b-8b61-776c3b41f77e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 568.066792] env[63021]: ERROR nova.compute.manager [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 60d66191-46ca-4e31-8dda-90f2a0c92b59, please check neutron logs for more information. [ 568.066792] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 568.066792] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.066792] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 568.066792] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 568.066792] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 568.066792] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 568.066792] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 568.066792] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.066792] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 568.066792] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.066792] env[63021]: ERROR nova.compute.manager raise self.value [ 568.066792] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 568.066792] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 568.066792] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.066792] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 568.067281] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.067281] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 568.067281] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 60d66191-46ca-4e31-8dda-90f2a0c92b59, please check neutron logs for more information. [ 568.067281] env[63021]: ERROR nova.compute.manager [ 568.067281] env[63021]: Traceback (most recent call last): [ 568.067281] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 568.067281] env[63021]: listener.cb(fileno) [ 568.067281] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 568.067281] env[63021]: result = function(*args, **kwargs) [ 568.067281] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 568.067281] env[63021]: return func(*args, **kwargs) [ 568.067281] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 568.067281] env[63021]: raise e [ 568.067281] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.067281] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 568.067281] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 568.067281] env[63021]: created_port_ids = self._update_ports_for_instance( [ 568.067281] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 568.067281] env[63021]: with excutils.save_and_reraise_exception(): [ 568.067281] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.067281] env[63021]: self.force_reraise() [ 568.067281] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.067281] env[63021]: raise self.value [ 568.067281] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 568.067281] env[63021]: updated_port = self._update_port( [ 568.067281] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.067281] env[63021]: _ensure_no_port_binding_failure(port) [ 568.067281] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.067281] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 568.067971] env[63021]: nova.exception.PortBindingFailed: Binding failed for port 60d66191-46ca-4e31-8dda-90f2a0c92b59, please check neutron logs for more information. [ 568.067971] env[63021]: Removing descriptor: 15 [ 568.067971] env[63021]: ERROR nova.compute.manager [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 60d66191-46ca-4e31-8dda-90f2a0c92b59, please check neutron logs for more information. [ 568.067971] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Traceback (most recent call last): [ 568.067971] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 568.067971] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] yield resources [ 568.067971] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 568.067971] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] self.driver.spawn(context, instance, image_meta, [ 568.067971] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 568.067971] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 568.067971] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 568.067971] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] vm_ref = self.build_virtual_machine(instance, [ 568.068324] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 568.068324] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] vif_infos = vmwarevif.get_vif_info(self._session, [ 568.068324] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 568.068324] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] for vif in network_info: [ 568.068324] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 568.068324] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] return self._sync_wrapper(fn, *args, **kwargs) [ 568.068324] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 568.068324] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] self.wait() [ 568.068324] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 568.068324] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] self[:] = self._gt.wait() [ 568.068324] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 568.068324] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] return self._exit_event.wait() [ 568.068324] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 568.068673] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] result = hub.switch() [ 568.068673] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 568.068673] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] return self.greenlet.switch() [ 568.068673] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 568.068673] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] result = function(*args, **kwargs) [ 568.068673] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 568.068673] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] return func(*args, **kwargs) [ 568.068673] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 568.068673] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] raise e [ 568.068673] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.068673] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] nwinfo = self.network_api.allocate_for_instance( [ 568.068673] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 568.068673] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] created_port_ids = self._update_ports_for_instance( [ 568.068992] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 568.068992] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] with excutils.save_and_reraise_exception(): [ 568.068992] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.068992] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] self.force_reraise() [ 568.068992] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.068992] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] raise self.value [ 568.068992] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 568.068992] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] updated_port = self._update_port( [ 568.068992] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.068992] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] _ensure_no_port_binding_failure(port) [ 568.068992] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.068992] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] raise exception.PortBindingFailed(port_id=port['id']) [ 568.069317] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] nova.exception.PortBindingFailed: Binding failed for port 60d66191-46ca-4e31-8dda-90f2a0c92b59, please check neutron logs for more information. [ 568.069317] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] [ 568.069317] env[63021]: INFO nova.compute.manager [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Terminating instance [ 568.072460] env[63021]: DEBUG oslo_concurrency.lockutils [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Acquiring lock "refresh_cache-2201799a-2df0-4fba-bcce-1fd8dfa1b6b9" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 568.072460] env[63021]: DEBUG oslo_concurrency.lockutils [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Acquired lock "refresh_cache-2201799a-2df0-4fba-bcce-1fd8dfa1b6b9" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 568.073034] env[63021]: DEBUG nova.network.neutron [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 568.140274] env[63021]: DEBUG oslo_vmware.api [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5242bd31-0da3-7ca4-4578-287e7bc36236, 'name': SearchDatastore_Task, 'duration_secs': 0.02048} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 568.140644] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 568.141611] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd/5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 568.141611] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-36f44d52-6772-4bc6-abf3-f3f82a52f27a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.150225] env[63021]: DEBUG oslo_vmware.api [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 568.150225] env[63021]: value = "task-1293530" [ 568.150225] env[63021]: _type = "Task" [ 568.150225] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 568.160461] env[63021]: DEBUG oslo_vmware.api [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293530, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 568.298357] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293529, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065959} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 568.302073] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 568.304031] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1658bfaf-388f-4a8f-be61-3207dfdc64d8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.339301] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Reconfiguring VM instance instance-00000008 to attach disk [datastore1] 0950cbd3-0708-4994-8d2d-053aecc06cdf/0950cbd3-0708-4994-8d2d-053aecc06cdf.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 568.340644] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dcb7f7c3-162c-4c7c-bc0c-926d0782671a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.365149] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 568.365149] env[63021]: value = "task-1293531" [ 568.365149] env[63021]: _type = "Task" [ 568.365149] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 568.376608] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293531, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 568.448953] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance efd0e375-3134-4984-8f82-48b53c65e26c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 568.449319] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=63021) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 568.449556] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=63021) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 568.627742] env[63021]: DEBUG nova.network.neutron [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 568.665420] env[63021]: DEBUG oslo_vmware.api [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293530, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 568.674597] env[63021]: DEBUG nova.compute.manager [req-45228282-8738-4d25-895e-ea6772fa68ca req-39e5c1a1-abbe-4c71-82a1-e32a8e7177d0 service nova] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Received event network-vif-deleted-f2cc2dc3-c7ef-4dcc-bbf3-c9154df16cc6 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 568.767159] env[63021]: DEBUG nova.network.neutron [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.880211] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293531, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 568.941853] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82881f28-b86f-42f9-bee1-17e90977f504 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.955931] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e975977c-540e-49fb-a533-127e84600bd0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.986187] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-575f7071-8b56-4b19-9d58-98891b1b22a5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.994486] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97fa326f-de07-4dc9-92de-6c487ad08bdd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.008540] env[63021]: DEBUG nova.compute.provider_tree [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 569.014024] env[63021]: DEBUG nova.compute.manager [req-8a6930ce-b834-4a59-b391-55ca0683f09c req-3118b82c-3917-485f-ac40-8fb04f8fd2da service nova] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Received event network-changed-60d66191-46ca-4e31-8dda-90f2a0c92b59 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 569.014466] env[63021]: DEBUG nova.compute.manager [req-8a6930ce-b834-4a59-b391-55ca0683f09c req-3118b82c-3917-485f-ac40-8fb04f8fd2da service nova] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Refreshing instance network info cache due to event network-changed-60d66191-46ca-4e31-8dda-90f2a0c92b59. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 569.014757] env[63021]: DEBUG oslo_concurrency.lockutils [req-8a6930ce-b834-4a59-b391-55ca0683f09c req-3118b82c-3917-485f-ac40-8fb04f8fd2da service nova] Acquiring lock "refresh_cache-2201799a-2df0-4fba-bcce-1fd8dfa1b6b9" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 569.088013] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.088013] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.163561] env[63021]: DEBUG oslo_vmware.api [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293530, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.583299} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 569.163857] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd/5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 569.164068] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 569.164275] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-72991347-a5c9-4f6f-9179-bb26364f772b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.173477] env[63021]: DEBUG oslo_vmware.api [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 569.173477] env[63021]: value = "task-1293532" [ 569.173477] env[63021]: _type = "Task" [ 569.173477] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 569.181780] env[63021]: DEBUG oslo_vmware.api [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293532, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 569.269311] env[63021]: DEBUG oslo_concurrency.lockutils [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Releasing lock "refresh_cache-2201799a-2df0-4fba-bcce-1fd8dfa1b6b9" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 569.269669] env[63021]: DEBUG nova.compute.manager [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 569.269861] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 569.270190] env[63021]: DEBUG oslo_concurrency.lockutils [req-8a6930ce-b834-4a59-b391-55ca0683f09c req-3118b82c-3917-485f-ac40-8fb04f8fd2da service nova] Acquired lock "refresh_cache-2201799a-2df0-4fba-bcce-1fd8dfa1b6b9" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 569.270348] env[63021]: DEBUG nova.network.neutron [req-8a6930ce-b834-4a59-b391-55ca0683f09c req-3118b82c-3917-485f-ac40-8fb04f8fd2da service nova] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Refreshing network info cache for port 60d66191-46ca-4e31-8dda-90f2a0c92b59 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 569.271479] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9572df66-7ba0-4400-91a1-0cfc2d8affb9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.279699] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a72c7bc-bf9d-4464-abb7-7c68b4510a69 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.302917] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9 could not be found. [ 569.303149] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 569.303323] env[63021]: INFO nova.compute.manager [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Took 0.03 seconds to destroy the instance on the hypervisor. [ 569.303556] env[63021]: DEBUG oslo.service.loopingcall [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 569.303758] env[63021]: DEBUG nova.compute.manager [-] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 569.303869] env[63021]: DEBUG nova.network.neutron [-] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 569.322901] env[63021]: DEBUG nova.network.neutron [-] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 569.378968] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293531, 'name': ReconfigVM_Task, 'duration_secs': 0.626651} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 569.382459] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Reconfigured VM instance instance-00000008 to attach disk [datastore1] 0950cbd3-0708-4994-8d2d-053aecc06cdf/0950cbd3-0708-4994-8d2d-053aecc06cdf.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 569.382459] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4769326a-98c5-494f-984e-343f3facb4ac {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.392061] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 569.392061] env[63021]: value = "task-1293533" [ 569.392061] env[63021]: _type = "Task" [ 569.392061] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 569.406650] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293533, 'name': Rename_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 569.514974] env[63021]: DEBUG nova.scheduler.client.report [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 569.685209] env[63021]: DEBUG oslo_vmware.api [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293532, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075139} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 569.685486] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 569.686345] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e03184c-9232-42a4-874a-f57231c0958f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.711063] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Reconfiguring VM instance instance-0000000a to attach disk [datastore1] 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd/5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 569.711159] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-df6f8e27-1e09-4627-a5fb-03e5d6177a3d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.730433] env[63021]: DEBUG oslo_vmware.api [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 569.730433] env[63021]: value = "task-1293534" [ 569.730433] env[63021]: _type = "Task" [ 569.730433] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 569.743055] env[63021]: DEBUG oslo_vmware.api [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293534, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 569.807182] env[63021]: DEBUG nova.network.neutron [req-8a6930ce-b834-4a59-b391-55ca0683f09c req-3118b82c-3917-485f-ac40-8fb04f8fd2da service nova] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 569.826375] env[63021]: DEBUG nova.network.neutron [-] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.906760] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293533, 'name': Rename_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 569.908051] env[63021]: DEBUG nova.network.neutron [req-8a6930ce-b834-4a59-b391-55ca0683f09c req-3118b82c-3917-485f-ac40-8fb04f8fd2da service nova] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.020754] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Acquiring lock "7da1d0f4-5dbe-4243-ac87-e476d5870882" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.021163] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Lock "7da1d0f4-5dbe-4243-ac87-e476d5870882" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.023340] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63021) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 570.023340] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 7.173s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.023408] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.788s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.243739] env[63021]: DEBUG oslo_vmware.api [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293534, 'name': ReconfigVM_Task, 'duration_secs': 0.318199} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 570.245159] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Reconfigured VM instance instance-0000000a to attach disk [datastore1] 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd/5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 570.248123] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7b82d488-27dd-4c27-b31e-961beb888ecb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.258029] env[63021]: DEBUG oslo_vmware.api [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 570.258029] env[63021]: value = "task-1293535" [ 570.258029] env[63021]: _type = "Task" [ 570.258029] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 570.265368] env[63021]: DEBUG oslo_vmware.api [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293535, 'name': Rename_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 570.331119] env[63021]: INFO nova.compute.manager [-] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Took 1.03 seconds to deallocate network for instance. [ 570.334209] env[63021]: DEBUG nova.compute.claims [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 570.334392] env[63021]: DEBUG oslo_concurrency.lockutils [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.404440] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293533, 'name': Rename_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 570.411238] env[63021]: DEBUG oslo_concurrency.lockutils [req-8a6930ce-b834-4a59-b391-55ca0683f09c req-3118b82c-3917-485f-ac40-8fb04f8fd2da service nova] Releasing lock "refresh_cache-2201799a-2df0-4fba-bcce-1fd8dfa1b6b9" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 570.768153] env[63021]: DEBUG oslo_vmware.api [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293535, 'name': Rename_Task, 'duration_secs': 0.158623} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 570.768386] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 570.768643] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1dbb774a-1517-4e6f-8847-1066169728fe {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.775904] env[63021]: DEBUG oslo_vmware.api [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 570.775904] env[63021]: value = "task-1293536" [ 570.775904] env[63021]: _type = "Task" [ 570.775904] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 570.786504] env[63021]: DEBUG oslo_vmware.api [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293536, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 570.907895] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293533, 'name': Rename_Task, 'duration_secs': 1.01829} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 570.908357] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 570.909468] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fa058a41-91f3-479f-ae98-bdc85f45cec5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.911852] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfdc67af-74a1-44f7-972b-d4433ab02e36 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.920485] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-513fdb90-5b5f-4908-9086-0f0060fbf9ef {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.923768] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 570.923768] env[63021]: value = "task-1293537" [ 570.923768] env[63021]: _type = "Task" [ 570.923768] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 570.954670] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eed48d7b-e0a5-48ae-ae04-d1465907fc8b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.961058] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293537, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 570.966198] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd479efd-cabe-41bd-8554-086e8800ae24 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.981412] env[63021]: DEBUG nova.compute.provider_tree [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 571.143940] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Acquiring lock "c76dfad0-2c67-4fe4-8c86-04ece5c80ea8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.144198] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Lock "c76dfad0-2c67-4fe4-8c86-04ece5c80ea8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.286140] env[63021]: DEBUG oslo_vmware.api [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293536, 'name': PowerOnVM_Task, 'duration_secs': 0.506808} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 571.286464] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 571.288276] env[63021]: INFO nova.compute.manager [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Took 6.41 seconds to spawn the instance on the hypervisor. [ 571.288276] env[63021]: DEBUG nova.compute.manager [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 571.288276] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f78285b7-ac51-4c76-b6ec-4186f3201b91 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.435799] env[63021]: DEBUG oslo_vmware.api [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293537, 'name': PowerOnVM_Task, 'duration_secs': 0.46955} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 571.435799] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 571.436236] env[63021]: INFO nova.compute.manager [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Took 11.35 seconds to spawn the instance on the hypervisor. [ 571.436236] env[63021]: DEBUG nova.compute.manager [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 571.437061] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfefba7a-da06-41cd-959c-ff07ac5785ab {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.484392] env[63021]: DEBUG nova.scheduler.client.report [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 571.821714] env[63021]: INFO nova.compute.manager [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Took 19.09 seconds to build instance. [ 571.953831] env[63021]: INFO nova.compute.manager [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Took 21.82 seconds to build instance. [ 571.989156] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.966s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 571.989792] env[63021]: ERROR nova.compute.manager [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a1a43b46-a34b-4ebe-befa-e72aa56aae02, please check neutron logs for more information. [ 571.989792] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Traceback (most recent call last): [ 571.989792] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 571.989792] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] self.driver.spawn(context, instance, image_meta, [ 571.989792] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 571.989792] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 571.989792] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 571.989792] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] vm_ref = self.build_virtual_machine(instance, [ 571.989792] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 571.989792] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] vif_infos = vmwarevif.get_vif_info(self._session, [ 571.989792] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 571.990135] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] for vif in network_info: [ 571.990135] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 571.990135] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] return self._sync_wrapper(fn, *args, **kwargs) [ 571.990135] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 571.990135] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] self.wait() [ 571.990135] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 571.990135] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] self[:] = self._gt.wait() [ 571.990135] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 571.990135] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] return self._exit_event.wait() [ 571.990135] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 571.990135] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] result = hub.switch() [ 571.990135] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 571.990135] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] return self.greenlet.switch() [ 571.990477] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 571.990477] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] result = function(*args, **kwargs) [ 571.990477] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 571.990477] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] return func(*args, **kwargs) [ 571.990477] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 571.990477] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] raise e [ 571.990477] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 571.990477] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] nwinfo = self.network_api.allocate_for_instance( [ 571.990477] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 571.990477] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] created_port_ids = self._update_ports_for_instance( [ 571.990477] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 571.990477] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] with excutils.save_and_reraise_exception(): [ 571.990477] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 571.990816] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] self.force_reraise() [ 571.990816] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 571.990816] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] raise self.value [ 571.990816] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 571.990816] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] updated_port = self._update_port( [ 571.990816] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 571.990816] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] _ensure_no_port_binding_failure(port) [ 571.990816] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 571.990816] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] raise exception.PortBindingFailed(port_id=port['id']) [ 571.990816] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] nova.exception.PortBindingFailed: Binding failed for port a1a43b46-a34b-4ebe-befa-e72aa56aae02, please check neutron logs for more information. [ 571.990816] env[63021]: ERROR nova.compute.manager [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] [ 571.991105] env[63021]: DEBUG nova.compute.utils [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Binding failed for port a1a43b46-a34b-4ebe-befa-e72aa56aae02, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 571.994502] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.427s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.994502] env[63021]: INFO nova.compute.claims [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 571.997745] env[63021]: DEBUG nova.compute.manager [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Build of instance d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f was re-scheduled: Binding failed for port a1a43b46-a34b-4ebe-befa-e72aa56aae02, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 571.998272] env[63021]: DEBUG nova.compute.manager [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 571.998564] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Acquiring lock "refresh_cache-d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 571.998801] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Acquired lock "refresh_cache-d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 571.999041] env[63021]: DEBUG nova.network.neutron [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 572.323905] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4085ebf0-6c7c-492e-9c3e-360cda8e1673 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Lock "5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.607s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.455660] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e0a1c95a-1d8f-44c9-9d28-b87629224b66 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Lock "0950cbd3-0708-4994-8d2d-053aecc06cdf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.340s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.516859] env[63021]: DEBUG nova.network.neutron [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 572.566085] env[63021]: DEBUG nova.network.neutron [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 572.588081] env[63021]: DEBUG nova.compute.manager [req-beb09ab8-c04b-4af9-bd1f-be256b0471de req-93234c02-bd1d-4760-85e6-8de16a1abf89 service nova] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Received event network-vif-deleted-60d66191-46ca-4e31-8dda-90f2a0c92b59 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 572.828357] env[63021]: DEBUG nova.compute.manager [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 572.960183] env[63021]: DEBUG nova.compute.manager [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 573.071027] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Releasing lock "refresh_cache-d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 573.071027] env[63021]: DEBUG nova.compute.manager [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 573.071027] env[63021]: DEBUG nova.compute.manager [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 573.071027] env[63021]: DEBUG nova.network.neutron [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 573.092337] env[63021]: DEBUG nova.network.neutron [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 573.352634] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.386872] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94bed61a-665c-4c7c-ac3d-200ad4e68c4c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.394690] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc77f547-2db2-484c-ba65-451a46a02172 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.426662] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc6af84d-bfe0-4f23-9618-63d3ab03a285 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.438352] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f95d919-b5d0-44be-9fca-4ee3aad0f982 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.453252] env[63021]: DEBUG nova.compute.provider_tree [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 573.491109] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.595220] env[63021]: DEBUG nova.network.neutron [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.956540] env[63021]: DEBUG nova.scheduler.client.report [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 574.021705] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Acquiring lock "1aaa84d5-c5d5-4321-a98f-da261de5d364" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.021881] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Lock "1aaa84d5-c5d5-4321-a98f-da261de5d364" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.028010] env[63021]: INFO nova.compute.manager [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Rebuilding instance [ 574.089613] env[63021]: DEBUG nova.compute.manager [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 574.090645] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b3ca746-00bf-464c-8be4-756894dfb1c9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.099639] env[63021]: INFO nova.compute.manager [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f] Took 1.03 seconds to deallocate network for instance. [ 574.465491] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.471s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 574.465491] env[63021]: DEBUG nova.compute.manager [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 574.467816] env[63021]: DEBUG oslo_concurrency.lockutils [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.395s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.610442] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 574.610697] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f27629b6-7131-48e8-bd4f-3e0f8ede88ad {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.619216] env[63021]: DEBUG oslo_vmware.api [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 574.619216] env[63021]: value = "task-1293542" [ 574.619216] env[63021]: _type = "Task" [ 574.619216] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 574.627994] env[63021]: DEBUG oslo_vmware.api [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293542, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 574.974808] env[63021]: DEBUG nova.compute.utils [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 574.983484] env[63021]: DEBUG nova.compute.manager [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Not allocating networking since 'none' was specified. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 575.130418] env[63021]: DEBUG oslo_vmware.api [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293542, 'name': PowerOffVM_Task, 'duration_secs': 0.189652} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 575.130524] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 575.130666] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 575.131766] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-386ff1aa-7f45-4441-9e0c-23a095c119db {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.144500] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 575.144500] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d3176fda-737f-4903-a7b7-d34535810526 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.147112] env[63021]: INFO nova.scheduler.client.report [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Deleted allocations for instance d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f [ 575.169851] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 575.169851] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 575.170674] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Deleting the datastore file [datastore1] 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 575.170674] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7c530284-130e-4eca-9049-697f16bd743f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.180621] env[63021]: DEBUG oslo_vmware.api [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 575.180621] env[63021]: value = "task-1293544" [ 575.180621] env[63021]: _type = "Task" [ 575.180621] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 575.200590] env[63021]: DEBUG oslo_vmware.api [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293544, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 575.425817] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba539398-032b-42e6-b100-7a04b52f08a4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.433650] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceadb2dc-c011-464b-a20b-562707395eb4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.465148] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0edb363d-6c83-45d2-a06a-e9443ba1c591 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.473692] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24200da2-4546-4798-8864-5b46f58460a7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.490826] env[63021]: DEBUG nova.compute.manager [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 575.492690] env[63021]: DEBUG nova.compute.provider_tree [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 575.668620] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0673a797-3960-4b35-b4d6-35fc11763c86 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Lock "d4ae25ce-04e5-41d1-8f3e-cda3e50cd97f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.115s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 575.727338] env[63021]: DEBUG oslo_vmware.api [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293544, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.308739} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 575.729048] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 575.729048] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 575.729048] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 576.001600] env[63021]: DEBUG nova.scheduler.client.report [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 576.171745] env[63021]: DEBUG nova.compute.manager [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 576.507332] env[63021]: DEBUG oslo_concurrency.lockutils [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.039s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 576.508284] env[63021]: ERROR nova.compute.manager [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 80b3a05d-0a47-4bb2-ae3b-41543c11af6c, please check neutron logs for more information. [ 576.508284] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Traceback (most recent call last): [ 576.508284] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 576.508284] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] self.driver.spawn(context, instance, image_meta, [ 576.508284] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 576.508284] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] self._vmops.spawn(context, instance, image_meta, injected_files, [ 576.508284] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 576.508284] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] vm_ref = self.build_virtual_machine(instance, [ 576.508284] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 576.508284] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] vif_infos = vmwarevif.get_vif_info(self._session, [ 576.508284] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 576.508669] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] for vif in network_info: [ 576.508669] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 576.508669] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] return self._sync_wrapper(fn, *args, **kwargs) [ 576.508669] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 576.508669] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] self.wait() [ 576.508669] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 576.508669] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] self[:] = self._gt.wait() [ 576.508669] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 576.508669] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] return self._exit_event.wait() [ 576.508669] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 576.508669] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] result = hub.switch() [ 576.508669] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 576.508669] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] return self.greenlet.switch() [ 576.509141] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 576.509141] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] result = function(*args, **kwargs) [ 576.509141] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 576.509141] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] return func(*args, **kwargs) [ 576.509141] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 576.509141] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] raise e [ 576.509141] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 576.509141] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] nwinfo = self.network_api.allocate_for_instance( [ 576.509141] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 576.509141] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] created_port_ids = self._update_ports_for_instance( [ 576.509141] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 576.509141] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] with excutils.save_and_reraise_exception(): [ 576.509141] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 576.509495] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] self.force_reraise() [ 576.509495] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 576.509495] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] raise self.value [ 576.509495] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 576.509495] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] updated_port = self._update_port( [ 576.509495] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 576.509495] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] _ensure_no_port_binding_failure(port) [ 576.509495] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 576.509495] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] raise exception.PortBindingFailed(port_id=port['id']) [ 576.509495] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] nova.exception.PortBindingFailed: Binding failed for port 80b3a05d-0a47-4bb2-ae3b-41543c11af6c, please check neutron logs for more information. [ 576.509495] env[63021]: ERROR nova.compute.manager [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] [ 576.510368] env[63021]: DEBUG nova.compute.utils [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Binding failed for port 80b3a05d-0a47-4bb2-ae3b-41543c11af6c, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 576.513169] env[63021]: DEBUG nova.compute.manager [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 576.519200] env[63021]: DEBUG nova.compute.manager [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Build of instance d629c773-3eb1-4f8d-9888-8bbf7565acea was re-scheduled: Binding failed for port 80b3a05d-0a47-4bb2-ae3b-41543c11af6c, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 576.519644] env[63021]: DEBUG nova.compute.manager [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 576.519883] env[63021]: DEBUG oslo_concurrency.lockutils [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Acquiring lock "refresh_cache-d629c773-3eb1-4f8d-9888-8bbf7565acea" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 576.520036] env[63021]: DEBUG oslo_concurrency.lockutils [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Acquired lock "refresh_cache-d629c773-3eb1-4f8d-9888-8bbf7565acea" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 576.520451] env[63021]: DEBUG nova.network.neutron [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 576.522566] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.232s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.524156] env[63021]: INFO nova.compute.claims [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 576.557637] env[63021]: DEBUG nova.virt.hardware [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 576.557878] env[63021]: DEBUG nova.virt.hardware [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 576.558051] env[63021]: DEBUG nova.virt.hardware [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 576.558236] env[63021]: DEBUG nova.virt.hardware [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 576.558545] env[63021]: DEBUG nova.virt.hardware [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 576.558545] env[63021]: DEBUG nova.virt.hardware [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 576.558760] env[63021]: DEBUG nova.virt.hardware [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 576.558912] env[63021]: DEBUG nova.virt.hardware [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 576.559093] env[63021]: DEBUG nova.virt.hardware [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 576.559257] env[63021]: DEBUG nova.virt.hardware [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 576.559425] env[63021]: DEBUG nova.virt.hardware [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 576.560366] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08c1ccc5-901d-47d4-943d-01fba70a8986 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.568822] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a27f9a1-03b5-419c-9902-8e09eba18626 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.584976] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Instance VIF info [] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 576.591072] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Creating folder: Project (45210d8f3be24f3c929c661905befd47). Parent ref: group-v277447. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 576.591473] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-68d1e4a7-8fdf-46ca-a0a3-f7d9044fd3b2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.603962] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Created folder: Project (45210d8f3be24f3c929c661905befd47) in parent group-v277447. [ 576.603962] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Creating folder: Instances. Parent ref: group-v277455. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 576.603962] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a6ca6b5d-4b04-4666-b21e-c0ffa1ffdc15 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.611700] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Created folder: Instances in parent group-v277455. [ 576.611970] env[63021]: DEBUG oslo.service.loopingcall [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 576.612185] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 576.612410] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a736b439-4c78-4088-92d7-a14de001a678 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.630019] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 576.630019] env[63021]: value = "task-1293548" [ 576.630019] env[63021]: _type = "Task" [ 576.630019] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 576.639722] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293548, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 576.698714] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.770506] env[63021]: DEBUG nova.virt.hardware [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 576.772026] env[63021]: DEBUG nova.virt.hardware [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 576.772026] env[63021]: DEBUG nova.virt.hardware [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 576.772026] env[63021]: DEBUG nova.virt.hardware [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 576.772026] env[63021]: DEBUG nova.virt.hardware [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 576.772026] env[63021]: DEBUG nova.virt.hardware [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 576.772026] env[63021]: DEBUG nova.virt.hardware [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 576.772266] env[63021]: DEBUG nova.virt.hardware [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 576.772266] env[63021]: DEBUG nova.virt.hardware [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 576.772266] env[63021]: DEBUG nova.virt.hardware [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 576.772360] env[63021]: DEBUG nova.virt.hardware [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 576.773461] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b45c78ca-8263-4e2d-b63d-48f1d70316a7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.782327] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90656766-7e23-4dca-af7e-be12e7203918 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.800229] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Instance VIF info [] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 576.807761] env[63021]: DEBUG oslo.service.loopingcall [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 576.808640] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 576.809156] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e9fef621-4e25-4f48-9331-669276825bad {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.831134] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 576.831134] env[63021]: value = "task-1293549" [ 576.831134] env[63021]: _type = "Task" [ 576.831134] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 576.839648] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293549, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.052751] env[63021]: DEBUG nova.network.neutron [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 577.128810] env[63021]: DEBUG nova.network.neutron [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.145833] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293548, 'name': CreateVM_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.342604] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293549, 'name': CreateVM_Task} progress is 25%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.632643] env[63021]: DEBUG oslo_concurrency.lockutils [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Releasing lock "refresh_cache-d629c773-3eb1-4f8d-9888-8bbf7565acea" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 577.633484] env[63021]: DEBUG nova.compute.manager [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 577.633725] env[63021]: DEBUG nova.compute.manager [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 577.633902] env[63021]: DEBUG nova.network.neutron [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 577.652569] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293548, 'name': CreateVM_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.662539] env[63021]: DEBUG nova.network.neutron [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 577.848220] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293549, 'name': CreateVM_Task} progress is 25%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 578.029959] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4b5794c-e24e-4fbf-968f-bf96ee1a0e1b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.038191] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b629a24f-2622-4049-828e-4b6a3b824b3c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.075309] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06d5cca7-a8f3-478c-9574-e41dcb092f79 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.082814] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d51a635-07c6-4975-abb5-d6ff034f1f33 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.104021] env[63021]: DEBUG nova.compute.provider_tree [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 578.145293] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293548, 'name': CreateVM_Task, 'duration_secs': 1.349243} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 578.145451] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 578.146299] env[63021]: DEBUG oslo_vmware.service [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f95b0f5-26f0-430f-bde6-2d08a7fb77f9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.152365] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 578.152530] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 578.152889] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 578.158299] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac7ac049-e303-4264-a5db-191800583ba1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.158299] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Waiting for the task: (returnval){ [ 578.158299] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52ee99ba-fd41-3d1b-15bf-1da3a70029e0" [ 578.158299] env[63021]: _type = "Task" [ 578.158299] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 578.168018] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52ee99ba-fd41-3d1b-15bf-1da3a70029e0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 578.168171] env[63021]: DEBUG nova.network.neutron [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.341368] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293549, 'name': CreateVM_Task, 'duration_secs': 1.317254} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 578.341560] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 578.342026] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 578.606072] env[63021]: DEBUG nova.scheduler.client.report [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 578.669710] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.669885] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 578.670128] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 578.670292] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 578.670492] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 578.670822] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 578.671214] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 578.671355] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-965ee793-0333-4eba-8acd-99874492dec6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.673427] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b33d0e3a-c649-4795-a419-cccb3efed126 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.675230] env[63021]: INFO nova.compute.manager [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] [instance: d629c773-3eb1-4f8d-9888-8bbf7565acea] Took 1.04 seconds to deallocate network for instance. [ 578.684968] env[63021]: DEBUG oslo_vmware.api [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 578.684968] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5227cc33-7bb3-38ef-7bce-421e7e03e198" [ 578.684968] env[63021]: _type = "Task" [ 578.684968] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 578.690950] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 578.691278] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 578.692396] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c59665a4-23d5-48d7-afb8-ef9d8f032a2e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.698229] env[63021]: DEBUG oslo_vmware.api [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5227cc33-7bb3-38ef-7bce-421e7e03e198, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 578.702532] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-46e133a5-f71e-4592-8746-729b92066d77 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.707411] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Waiting for the task: (returnval){ [ 578.707411] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]523f3d71-9487-b55d-0b30-7a3792980c4e" [ 578.707411] env[63021]: _type = "Task" [ 578.707411] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 578.715142] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]523f3d71-9487-b55d-0b30-7a3792980c4e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 579.113326] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.590s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.114344] env[63021]: DEBUG nova.compute.manager [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 579.126168] env[63021]: DEBUG oslo_concurrency.lockutils [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.074s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.198818] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 579.201015] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 579.201015] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.228018] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Preparing fetch location {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 579.228018] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Creating directory with path [datastore2] vmware_temp/8c995bd0-7d12-4ca1-bf96-a6d3a430456b/e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 579.228018] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7882003f-2d9a-4283-9d02-a47801930a80 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.263885] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Created directory with path [datastore2] vmware_temp/8c995bd0-7d12-4ca1-bf96-a6d3a430456b/e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 579.264590] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Fetch image to [datastore2] vmware_temp/8c995bd0-7d12-4ca1-bf96-a6d3a430456b/e624225d-af0f-415c-90cc-d4b228c479f6/tmp-sparse.vmdk {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 579.264796] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Downloading image file data e624225d-af0f-415c-90cc-d4b228c479f6 to [datastore2] vmware_temp/8c995bd0-7d12-4ca1-bf96-a6d3a430456b/e624225d-af0f-415c-90cc-d4b228c479f6/tmp-sparse.vmdk on the data store datastore2 {{(pid=63021) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 579.266546] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-244a1093-90ef-401a-9bc5-d102078a08a4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.277742] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcc28b5f-ace4-40bc-911a-02d2ed533bd1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.289457] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e6ff46a-df00-4c90-aa39-31d5cae39853 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.322094] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9129cad-b1af-43ed-8218-353e755315ba {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.328915] env[63021]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-4e94a40d-0750-4d92-9ec3-0ea4013e2f50 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.418651] env[63021]: DEBUG nova.virt.vmwareapi.images [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Downloading image file data e624225d-af0f-415c-90cc-d4b228c479f6 to the data store datastore2 {{(pid=63021) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 579.498063] env[63021]: DEBUG oslo_vmware.rw_handles [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/8c995bd0-7d12-4ca1-bf96-a6d3a430456b/e624225d-af0f-415c-90cc-d4b228c479f6/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=63021) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 579.638276] env[63021]: DEBUG nova.compute.utils [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 579.643411] env[63021]: DEBUG nova.compute.manager [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 579.643531] env[63021]: DEBUG nova.network.neutron [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 579.708665] env[63021]: INFO nova.scheduler.client.report [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Deleted allocations for instance d629c773-3eb1-4f8d-9888-8bbf7565acea [ 579.845074] env[63021]: DEBUG nova.policy [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '50d5450d7d1541388321de2146612925', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '44ac682161f24e29a7592fac57e1443b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 580.027767] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d7903e0-6d4d-4b0c-b264-d7656903eccf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.036342] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1651bf4d-c1ee-4876-b8b7-d4d511e96d32 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.075634] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-001ab413-77d7-499e-bc40-7e8120c35522 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.086864] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec58d499-f5e8-445f-aa09-a30086bce678 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.103187] env[63021]: DEBUG nova.compute.provider_tree [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 580.143740] env[63021]: DEBUG nova.compute.manager [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 580.173600] env[63021]: DEBUG oslo_vmware.rw_handles [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Completed reading data from the image iterator. {{(pid=63021) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 580.173600] env[63021]: DEBUG oslo_vmware.rw_handles [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Closing write handle for https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/8c995bd0-7d12-4ca1-bf96-a6d3a430456b/e624225d-af0f-415c-90cc-d4b228c479f6/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=63021) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 580.223845] env[63021]: DEBUG oslo_concurrency.lockutils [None req-174b31f9-6f9d-403c-8e1b-4f022fd9ff4f tempest-ServersTestJSON-1638932574 tempest-ServersTestJSON-1638932574-project-member] Lock "d629c773-3eb1-4f8d-9888-8bbf7565acea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.003s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.321690] env[63021]: DEBUG nova.virt.vmwareapi.images [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Downloaded image file data e624225d-af0f-415c-90cc-d4b228c479f6 to vmware_temp/8c995bd0-7d12-4ca1-bf96-a6d3a430456b/e624225d-af0f-415c-90cc-d4b228c479f6/tmp-sparse.vmdk on the data store datastore2 {{(pid=63021) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 580.323705] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Caching image {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 580.324205] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Copying Virtual Disk [datastore2] vmware_temp/8c995bd0-7d12-4ca1-bf96-a6d3a430456b/e624225d-af0f-415c-90cc-d4b228c479f6/tmp-sparse.vmdk to [datastore2] vmware_temp/8c995bd0-7d12-4ca1-bf96-a6d3a430456b/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 580.325527] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3fceafee-ba8a-41c9-9d47-1526d44656ad {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.340020] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Waiting for the task: (returnval){ [ 580.340020] env[63021]: value = "task-1293552" [ 580.340020] env[63021]: _type = "Task" [ 580.340020] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 580.349906] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Task: {'id': task-1293552, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 580.612103] env[63021]: DEBUG nova.scheduler.client.report [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 580.726962] env[63021]: DEBUG nova.compute.manager [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 580.751086] env[63021]: DEBUG nova.network.neutron [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Successfully created port: 40185534-7f2c-4524-98aa-8dad833d2c43 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 580.851930] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Task: {'id': task-1293552, 'name': CopyVirtualDisk_Task} progress is 27%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 581.125513] env[63021]: DEBUG oslo_concurrency.lockutils [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 581.125513] env[63021]: ERROR nova.compute.manager [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b1544ac9-c61e-4f0a-a311-99a23b19e457, please check neutron logs for more information. [ 581.125513] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Traceback (most recent call last): [ 581.125513] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 581.125513] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] self.driver.spawn(context, instance, image_meta, [ 581.125513] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 581.125513] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 581.125513] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 581.125513] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] vm_ref = self.build_virtual_machine(instance, [ 581.125882] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 581.125882] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] vif_infos = vmwarevif.get_vif_info(self._session, [ 581.125882] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 581.125882] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] for vif in network_info: [ 581.125882] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 581.125882] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] return self._sync_wrapper(fn, *args, **kwargs) [ 581.125882] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 581.125882] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] self.wait() [ 581.125882] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 581.125882] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] self[:] = self._gt.wait() [ 581.125882] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 581.125882] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] return self._exit_event.wait() [ 581.125882] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 581.126322] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] result = hub.switch() [ 581.126322] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 581.126322] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] return self.greenlet.switch() [ 581.126322] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 581.126322] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] result = function(*args, **kwargs) [ 581.126322] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 581.126322] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] return func(*args, **kwargs) [ 581.126322] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 581.126322] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] raise e [ 581.126322] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 581.126322] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] nwinfo = self.network_api.allocate_for_instance( [ 581.126322] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 581.126322] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] created_port_ids = self._update_ports_for_instance( [ 581.126732] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 581.126732] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] with excutils.save_and_reraise_exception(): [ 581.126732] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 581.126732] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] self.force_reraise() [ 581.126732] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 581.126732] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] raise self.value [ 581.126732] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 581.126732] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] updated_port = self._update_port( [ 581.126732] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 581.126732] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] _ensure_no_port_binding_failure(port) [ 581.126732] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 581.126732] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] raise exception.PortBindingFailed(port_id=port['id']) [ 581.128139] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] nova.exception.PortBindingFailed: Binding failed for port b1544ac9-c61e-4f0a-a311-99a23b19e457, please check neutron logs for more information. [ 581.128139] env[63021]: ERROR nova.compute.manager [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] [ 581.130857] env[63021]: DEBUG nova.compute.utils [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Binding failed for port b1544ac9-c61e-4f0a-a311-99a23b19e457, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 581.134377] env[63021]: DEBUG nova.compute.manager [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Build of instance 0b443010-4dcf-4a7a-aec3-2d50bd50acc3 was re-scheduled: Binding failed for port b1544ac9-c61e-4f0a-a311-99a23b19e457, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 581.135553] env[63021]: DEBUG nova.compute.manager [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 581.135937] env[63021]: DEBUG oslo_concurrency.lockutils [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Acquiring lock "refresh_cache-0b443010-4dcf-4a7a-aec3-2d50bd50acc3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 581.136206] env[63021]: DEBUG oslo_concurrency.lockutils [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Acquired lock "refresh_cache-0b443010-4dcf-4a7a-aec3-2d50bd50acc3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 581.136486] env[63021]: DEBUG nova.network.neutron [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 581.139441] env[63021]: DEBUG oslo_concurrency.lockutils [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.104s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 581.158964] env[63021]: DEBUG nova.compute.manager [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 581.198852] env[63021]: DEBUG nova.virt.hardware [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 581.198852] env[63021]: DEBUG nova.virt.hardware [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 581.198852] env[63021]: DEBUG nova.virt.hardware [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 581.202581] env[63021]: DEBUG nova.virt.hardware [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 581.203131] env[63021]: DEBUG nova.virt.hardware [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 581.203596] env[63021]: DEBUG nova.virt.hardware [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 581.204017] env[63021]: DEBUG nova.virt.hardware [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 581.204287] env[63021]: DEBUG nova.virt.hardware [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 581.204727] env[63021]: DEBUG nova.virt.hardware [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 581.205032] env[63021]: DEBUG nova.virt.hardware [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 581.205347] env[63021]: DEBUG nova.virt.hardware [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 581.207265] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-279a161a-7376-42e7-a583-c28fba12702d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.226214] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53083bc3-3091-4296-859e-a5d5f52daed1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.282490] env[63021]: DEBUG oslo_concurrency.lockutils [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.358314] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Task: {'id': task-1293552, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.96882} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 581.358314] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Copied Virtual Disk [datastore2] vmware_temp/8c995bd0-7d12-4ca1-bf96-a6d3a430456b/e624225d-af0f-415c-90cc-d4b228c479f6/tmp-sparse.vmdk to [datastore2] vmware_temp/8c995bd0-7d12-4ca1-bf96-a6d3a430456b/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 581.358314] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Deleting the datastore file [datastore2] vmware_temp/8c995bd0-7d12-4ca1-bf96-a6d3a430456b/e624225d-af0f-415c-90cc-d4b228c479f6/tmp-sparse.vmdk {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 581.358314] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c89e84d7-9229-4267-8695-3f9ace3329a5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.363130] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Waiting for the task: (returnval){ [ 581.363130] env[63021]: value = "task-1293553" [ 581.363130] env[63021]: _type = "Task" [ 581.363130] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 581.375849] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Task: {'id': task-1293553, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 581.699363] env[63021]: DEBUG nova.network.neutron [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 581.801212] env[63021]: DEBUG oslo_concurrency.lockutils [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Acquiring lock "f5f546f8-291f-430c-95fb-c78507b0087f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.801477] env[63021]: DEBUG oslo_concurrency.lockutils [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Lock "f5f546f8-291f-430c-95fb-c78507b0087f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 581.875836] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Task: {'id': task-1293553, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.021904} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 581.882552] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 581.882958] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Moving file from [datastore2] vmware_temp/8c995bd0-7d12-4ca1-bf96-a6d3a430456b/e624225d-af0f-415c-90cc-d4b228c479f6 to [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6. {{(pid=63021) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 581.885892] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-074ed624-59dd-4e48-ad1d-1b247080ab87 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.893326] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Waiting for the task: (returnval){ [ 581.893326] env[63021]: value = "task-1293554" [ 581.893326] env[63021]: _type = "Task" [ 581.893326] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 581.909334] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Task: {'id': task-1293554, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 582.183387] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e320bae7-3626-465c-933e-1f2bea1d3f94 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.192103] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f22c8471-ed74-4f2b-9587-45cb5da05a60 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.224340] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83469408-e22a-4b21-9584-119b10eeda3a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.231530] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9a5783e-eb58-4622-bfd6-57b174be0a72 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.249913] env[63021]: DEBUG nova.compute.provider_tree [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Updating inventory in ProviderTree for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 582.314223] env[63021]: DEBUG nova.network.neutron [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.404284] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Task: {'id': task-1293554, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.023496} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 582.405439] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] File moved {{(pid=63021) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 582.405439] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Cleaning up location [datastore2] vmware_temp/8c995bd0-7d12-4ca1-bf96-a6d3a430456b {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 582.405439] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Deleting the datastore file [datastore2] vmware_temp/8c995bd0-7d12-4ca1-bf96-a6d3a430456b {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 582.405439] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c23aa247-1bcb-4f89-8203-e8906127d830 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.415161] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Waiting for the task: (returnval){ [ 582.415161] env[63021]: value = "task-1293556" [ 582.415161] env[63021]: _type = "Task" [ 582.415161] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.425292] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Task: {'id': task-1293556, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 582.777518] env[63021]: ERROR nova.scheduler.client.report [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] [req-266bb5b7-d2a7-423f-9392-7588d43159a6] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 6ef6c041-2b0d-4bc3-8662-5d679a77a86d. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-266bb5b7-d2a7-423f-9392-7588d43159a6"}]}: nova.exception.PortBindingFailed: Binding failed for port 5adbb74f-e77e-4614-8abf-14b72f2f4658, please check neutron logs for more information. [ 582.799402] env[63021]: DEBUG nova.scheduler.client.report [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Refreshing inventories for resource provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 582.818243] env[63021]: DEBUG oslo_concurrency.lockutils [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Releasing lock "refresh_cache-0b443010-4dcf-4a7a-aec3-2d50bd50acc3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 582.818243] env[63021]: DEBUG nova.compute.manager [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 582.818243] env[63021]: DEBUG nova.compute.manager [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 582.818243] env[63021]: DEBUG nova.network.neutron [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 582.828041] env[63021]: DEBUG nova.scheduler.client.report [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Updating ProviderTree inventory for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 582.828245] env[63021]: DEBUG nova.compute.provider_tree [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Updating inventory in ProviderTree for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 582.845965] env[63021]: DEBUG nova.scheduler.client.report [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Refreshing aggregate associations for resource provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d, aggregates: None {{(pid=63021) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 582.853219] env[63021]: DEBUG nova.network.neutron [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 582.883918] env[63021]: DEBUG nova.scheduler.client.report [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Refreshing trait associations for resource provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE {{(pid=63021) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 582.928413] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Task: {'id': task-1293556, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.046666} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 582.929088] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 582.931485] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e51f481-a022-4a94-90f9-c2890b022625 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.943036] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Waiting for the task: (returnval){ [ 582.943036] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52a3d105-1418-63f5-5c1c-314df68e4a38" [ 582.943036] env[63021]: _type = "Task" [ 582.943036] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.954991] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52a3d105-1418-63f5-5c1c-314df68e4a38, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 583.192126] env[63021]: ERROR nova.compute.manager [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 40185534-7f2c-4524-98aa-8dad833d2c43, please check neutron logs for more information. [ 583.192126] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 583.192126] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 583.192126] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 583.192126] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 583.192126] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 583.192126] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 583.192126] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 583.192126] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 583.192126] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 583.192126] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 583.192126] env[63021]: ERROR nova.compute.manager raise self.value [ 583.192126] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 583.192126] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 583.192126] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 583.192126] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 583.192800] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 583.192800] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 583.192800] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 40185534-7f2c-4524-98aa-8dad833d2c43, please check neutron logs for more information. [ 583.192800] env[63021]: ERROR nova.compute.manager [ 583.192800] env[63021]: Traceback (most recent call last): [ 583.192800] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 583.192800] env[63021]: listener.cb(fileno) [ 583.192800] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 583.192800] env[63021]: result = function(*args, **kwargs) [ 583.192800] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 583.192800] env[63021]: return func(*args, **kwargs) [ 583.192800] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 583.192800] env[63021]: raise e [ 583.192800] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 583.192800] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 583.192800] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 583.192800] env[63021]: created_port_ids = self._update_ports_for_instance( [ 583.192800] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 583.192800] env[63021]: with excutils.save_and_reraise_exception(): [ 583.192800] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 583.192800] env[63021]: self.force_reraise() [ 583.192800] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 583.192800] env[63021]: raise self.value [ 583.192800] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 583.192800] env[63021]: updated_port = self._update_port( [ 583.192800] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 583.192800] env[63021]: _ensure_no_port_binding_failure(port) [ 583.192800] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 583.192800] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 583.193612] env[63021]: nova.exception.PortBindingFailed: Binding failed for port 40185534-7f2c-4524-98aa-8dad833d2c43, please check neutron logs for more information. [ 583.193612] env[63021]: Removing descriptor: 15 [ 583.199576] env[63021]: ERROR nova.compute.manager [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 40185534-7f2c-4524-98aa-8dad833d2c43, please check neutron logs for more information. [ 583.199576] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Traceback (most recent call last): [ 583.199576] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 583.199576] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] yield resources [ 583.199576] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 583.199576] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] self.driver.spawn(context, instance, image_meta, [ 583.199576] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 583.199576] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 583.199576] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 583.199576] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] vm_ref = self.build_virtual_machine(instance, [ 583.199576] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 583.200488] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] vif_infos = vmwarevif.get_vif_info(self._session, [ 583.200488] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 583.200488] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] for vif in network_info: [ 583.200488] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 583.200488] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] return self._sync_wrapper(fn, *args, **kwargs) [ 583.200488] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 583.200488] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] self.wait() [ 583.200488] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 583.200488] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] self[:] = self._gt.wait() [ 583.200488] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 583.200488] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] return self._exit_event.wait() [ 583.200488] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 583.200488] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] result = hub.switch() [ 583.200901] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 583.200901] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] return self.greenlet.switch() [ 583.200901] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 583.200901] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] result = function(*args, **kwargs) [ 583.200901] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 583.200901] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] return func(*args, **kwargs) [ 583.200901] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 583.200901] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] raise e [ 583.200901] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 583.200901] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] nwinfo = self.network_api.allocate_for_instance( [ 583.200901] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 583.200901] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] created_port_ids = self._update_ports_for_instance( [ 583.200901] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 583.202069] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] with excutils.save_and_reraise_exception(): [ 583.202069] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 583.202069] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] self.force_reraise() [ 583.202069] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 583.202069] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] raise self.value [ 583.202069] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 583.202069] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] updated_port = self._update_port( [ 583.202069] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 583.202069] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] _ensure_no_port_binding_failure(port) [ 583.202069] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 583.202069] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] raise exception.PortBindingFailed(port_id=port['id']) [ 583.202069] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] nova.exception.PortBindingFailed: Binding failed for port 40185534-7f2c-4524-98aa-8dad833d2c43, please check neutron logs for more information. [ 583.202069] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] [ 583.202545] env[63021]: INFO nova.compute.manager [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Terminating instance [ 583.202545] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Acquiring lock "refresh_cache-ba8e47f7-e10d-439a-9a5a-714655a29dc0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 583.202545] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Acquired lock "refresh_cache-ba8e47f7-e10d-439a-9a5a-714655a29dc0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.202545] env[63021]: DEBUG nova.network.neutron [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 583.356571] env[63021]: DEBUG nova.network.neutron [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.360031] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6f4bfde-5faa-4e82-95ae-072758cbc723 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.372755] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-006a1f25-a685-44f2-9ebd-0f41738e71c9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.419133] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8901427-d1db-4fb2-b0c0-fdfa2dec3b87 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.428462] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c800aa0f-f9e2-49d3-8f4b-2e62d0dd029c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.445373] env[63021]: DEBUG nova.compute.provider_tree [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Updating inventory in ProviderTree for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 583.457229] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52a3d105-1418-63f5-5c1c-314df68e4a38, 'name': SearchDatastore_Task, 'duration_secs': 0.009783} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 583.458181] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.458365] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 9d474bd0-e825-49f2-9715-14d85495718e/9d474bd0-e825-49f2-9715-14d85495718e.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 583.460766] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.460766] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 583.460766] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-af65b6d7-b1ac-4d81-b1a9-ca3f4971e5de {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.461278] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f66e4b0a-001a-4182-b6e3-80186c46cd58 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.467955] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Waiting for the task: (returnval){ [ 583.467955] env[63021]: value = "task-1293557" [ 583.467955] env[63021]: _type = "Task" [ 583.467955] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 583.472159] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 583.472332] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 583.473352] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9cb34795-a979-45ac-92aa-f91c71f79523 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.478500] env[63021]: DEBUG oslo_vmware.api [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 583.478500] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5268edf9-8560-3645-53c6-9fb4e6fe9a2b" [ 583.478500] env[63021]: _type = "Task" [ 583.478500] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 583.481704] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Task: {'id': task-1293557, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 583.489109] env[63021]: DEBUG oslo_vmware.api [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5268edf9-8560-3645-53c6-9fb4e6fe9a2b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 583.516802] env[63021]: DEBUG nova.compute.manager [req-845110fd-3f21-4436-b8be-e06ae6b5ac1f req-416949f3-0fc2-495c-a4d1-cde23e03ecc8 service nova] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Received event network-changed-40185534-7f2c-4524-98aa-8dad833d2c43 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 583.517285] env[63021]: DEBUG nova.compute.manager [req-845110fd-3f21-4436-b8be-e06ae6b5ac1f req-416949f3-0fc2-495c-a4d1-cde23e03ecc8 service nova] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Refreshing instance network info cache due to event network-changed-40185534-7f2c-4524-98aa-8dad833d2c43. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 583.518116] env[63021]: DEBUG oslo_concurrency.lockutils [req-845110fd-3f21-4436-b8be-e06ae6b5ac1f req-416949f3-0fc2-495c-a4d1-cde23e03ecc8 service nova] Acquiring lock "refresh_cache-ba8e47f7-e10d-439a-9a5a-714655a29dc0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 583.743847] env[63021]: DEBUG nova.network.neutron [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 583.865322] env[63021]: INFO nova.compute.manager [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] [instance: 0b443010-4dcf-4a7a-aec3-2d50bd50acc3] Took 1.05 seconds to deallocate network for instance. [ 583.899382] env[63021]: DEBUG nova.network.neutron [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.978112] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Task: {'id': task-1293557, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 584.000087] env[63021]: DEBUG oslo_vmware.api [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5268edf9-8560-3645-53c6-9fb4e6fe9a2b, 'name': SearchDatastore_Task, 'duration_secs': 0.01035} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 584.000359] env[63021]: DEBUG nova.scheduler.client.report [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Updated inventory for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with generation 33 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 584.000598] env[63021]: DEBUG nova.compute.provider_tree [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Updating resource provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d generation from 33 to 34 during operation: update_inventory {{(pid=63021) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 584.001842] env[63021]: DEBUG nova.compute.provider_tree [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Updating inventory in ProviderTree for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 584.008030] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e770362d-8c6e-4764-acf1-667f64486cfa {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.014442] env[63021]: DEBUG oslo_vmware.api [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 584.014442] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]526814b8-f437-4d0b-2ab7-28b6e604f5d9" [ 584.014442] env[63021]: _type = "Task" [ 584.014442] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 584.030873] env[63021]: DEBUG oslo_vmware.api [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]526814b8-f437-4d0b-2ab7-28b6e604f5d9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 584.408343] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Releasing lock "refresh_cache-ba8e47f7-e10d-439a-9a5a-714655a29dc0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 584.408994] env[63021]: DEBUG nova.compute.manager [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 584.413161] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 584.413161] env[63021]: DEBUG oslo_concurrency.lockutils [req-845110fd-3f21-4436-b8be-e06ae6b5ac1f req-416949f3-0fc2-495c-a4d1-cde23e03ecc8 service nova] Acquired lock "refresh_cache-ba8e47f7-e10d-439a-9a5a-714655a29dc0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 584.413161] env[63021]: DEBUG nova.network.neutron [req-845110fd-3f21-4436-b8be-e06ae6b5ac1f req-416949f3-0fc2-495c-a4d1-cde23e03ecc8 service nova] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Refreshing network info cache for port 40185534-7f2c-4524-98aa-8dad833d2c43 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 584.413161] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5998b072-9ca6-47d3-8316-73a5a26969b3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.425228] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ad1a10e-0cd0-4399-8f2f-5fa08238a99e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.453339] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ba8e47f7-e10d-439a-9a5a-714655a29dc0 could not be found. [ 584.453580] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 584.453784] env[63021]: INFO nova.compute.manager [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Took 0.04 seconds to destroy the instance on the hypervisor. [ 584.454052] env[63021]: DEBUG oslo.service.loopingcall [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 584.454307] env[63021]: DEBUG nova.compute.manager [-] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 584.454394] env[63021]: DEBUG nova.network.neutron [-] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 584.480304] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Task: {'id': task-1293557, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 584.505220] env[63021]: DEBUG nova.network.neutron [-] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 584.506946] env[63021]: DEBUG oslo_concurrency.lockutils [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 3.368s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 584.507550] env[63021]: ERROR nova.compute.manager [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5adbb74f-e77e-4614-8abf-14b72f2f4658, please check neutron logs for more information. [ 584.507550] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Traceback (most recent call last): [ 584.507550] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 584.507550] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] self.driver.spawn(context, instance, image_meta, [ 584.507550] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 584.507550] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] self._vmops.spawn(context, instance, image_meta, injected_files, [ 584.507550] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 584.507550] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] vm_ref = self.build_virtual_machine(instance, [ 584.507550] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 584.507550] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] vif_infos = vmwarevif.get_vif_info(self._session, [ 584.507550] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 584.507873] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] for vif in network_info: [ 584.507873] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 584.507873] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] return self._sync_wrapper(fn, *args, **kwargs) [ 584.507873] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 584.507873] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] self.wait() [ 584.507873] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 584.507873] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] self[:] = self._gt.wait() [ 584.507873] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 584.507873] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] return self._exit_event.wait() [ 584.507873] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 584.507873] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] result = hub.switch() [ 584.507873] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 584.507873] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] return self.greenlet.switch() [ 584.508236] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 584.508236] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] result = function(*args, **kwargs) [ 584.508236] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 584.508236] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] return func(*args, **kwargs) [ 584.508236] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 584.508236] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] raise e [ 584.508236] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 584.508236] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] nwinfo = self.network_api.allocate_for_instance( [ 584.508236] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 584.508236] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] created_port_ids = self._update_ports_for_instance( [ 584.508236] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 584.508236] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] with excutils.save_and_reraise_exception(): [ 584.508236] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.508732] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] self.force_reraise() [ 584.508732] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.508732] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] raise self.value [ 584.508732] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 584.508732] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] updated_port = self._update_port( [ 584.508732] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.508732] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] _ensure_no_port_binding_failure(port) [ 584.508732] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.508732] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] raise exception.PortBindingFailed(port_id=port['id']) [ 584.508732] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] nova.exception.PortBindingFailed: Binding failed for port 5adbb74f-e77e-4614-8abf-14b72f2f4658, please check neutron logs for more information. [ 584.508732] env[63021]: ERROR nova.compute.manager [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] [ 584.509050] env[63021]: DEBUG nova.compute.utils [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Binding failed for port 5adbb74f-e77e-4614-8abf-14b72f2f4658, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 584.512543] env[63021]: DEBUG nova.compute.manager [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Build of instance 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af was re-scheduled: Binding failed for port 5adbb74f-e77e-4614-8abf-14b72f2f4658, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 584.512543] env[63021]: DEBUG nova.compute.manager [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 584.512543] env[63021]: DEBUG oslo_concurrency.lockutils [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Acquiring lock "refresh_cache-7d98e2e4-33bc-4e6b-af6f-31c18f34b9af" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 584.512543] env[63021]: DEBUG oslo_concurrency.lockutils [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Acquired lock "refresh_cache-7d98e2e4-33bc-4e6b-af6f-31c18f34b9af" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 584.512693] env[63021]: DEBUG nova.network.neutron [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 584.512693] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.566s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.525507] env[63021]: DEBUG oslo_vmware.api [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]526814b8-f437-4d0b-2ab7-28b6e604f5d9, 'name': SearchDatastore_Task, 'duration_secs': 0.013201} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 584.526338] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 584.526594] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd/5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 584.526841] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2ac65519-0199-4b7c-92d8-addce277f72d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.534167] env[63021]: DEBUG oslo_vmware.api [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 584.534167] env[63021]: value = "task-1293559" [ 584.534167] env[63021]: _type = "Task" [ 584.534167] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 584.543843] env[63021]: DEBUG oslo_vmware.api [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293559, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 584.909577] env[63021]: INFO nova.scheduler.client.report [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Deleted allocations for instance 0b443010-4dcf-4a7a-aec3-2d50bd50acc3 [ 584.944985] env[63021]: DEBUG nova.network.neutron [req-845110fd-3f21-4436-b8be-e06ae6b5ac1f req-416949f3-0fc2-495c-a4d1-cde23e03ecc8 service nova] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 584.984654] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Task: {'id': task-1293557, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 585.013633] env[63021]: DEBUG nova.network.neutron [-] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.048915] env[63021]: DEBUG oslo_vmware.api [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293559, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 585.057485] env[63021]: DEBUG nova.network.neutron [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 585.193096] env[63021]: DEBUG nova.network.neutron [req-845110fd-3f21-4436-b8be-e06ae6b5ac1f req-416949f3-0fc2-495c-a4d1-cde23e03ecc8 service nova] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.309729] env[63021]: DEBUG nova.network.neutron [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.420558] env[63021]: DEBUG oslo_concurrency.lockutils [None req-43819cd4-6967-4a87-a42a-5ac8a0b4155a tempest-ServerAddressesTestJSON-966383762 tempest-ServerAddressesTestJSON-966383762-project-member] Lock "0b443010-4dcf-4a7a-aec3-2d50bd50acc3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.981s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.485134] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Task: {'id': task-1293557, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.547115} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 585.488322] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 9d474bd0-e825-49f2-9715-14d85495718e/9d474bd0-e825-49f2-9715-14d85495718e.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 585.489036] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 585.491617] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-308c1fb1-bfb2-4446-ba17-d36654f1e45d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.494152] env[63021]: DEBUG oslo_concurrency.lockutils [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Acquiring lock "05caabab-8c7d-4d27-8eb3-e7e1c2947aa6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.496018] env[63021]: DEBUG oslo_concurrency.lockutils [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Lock "05caabab-8c7d-4d27-8eb3-e7e1c2947aa6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.499506] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Waiting for the task: (returnval){ [ 585.499506] env[63021]: value = "task-1293560" [ 585.499506] env[63021]: _type = "Task" [ 585.499506] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 585.509274] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Task: {'id': task-1293560, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 585.516956] env[63021]: INFO nova.compute.manager [-] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Took 1.06 seconds to deallocate network for instance. [ 585.520082] env[63021]: DEBUG nova.compute.claims [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 585.520082] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.542026] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-832cc0d2-1cf1-46c9-910e-f2570c381346 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.552628] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ba44b5f-3835-405d-b47e-1ac7ddeab8fc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.556022] env[63021]: DEBUG oslo_vmware.api [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293559, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.964413} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 585.556261] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd/5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 585.556471] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 585.557079] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4f428166-e7ee-4478-8071-02868bccc8d2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.589290] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bdffc70-ddb4-4dad-9b15-a00743ed9b8d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.597085] env[63021]: DEBUG oslo_vmware.api [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 585.597085] env[63021]: value = "task-1293561" [ 585.597085] env[63021]: _type = "Task" [ 585.597085] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 585.611142] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a95925e7-fac4-427f-949f-d797c1cf2310 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.617984] env[63021]: DEBUG oslo_vmware.api [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293561, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 585.629573] env[63021]: DEBUG nova.compute.provider_tree [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 585.696828] env[63021]: DEBUG oslo_concurrency.lockutils [req-845110fd-3f21-4436-b8be-e06ae6b5ac1f req-416949f3-0fc2-495c-a4d1-cde23e03ecc8 service nova] Releasing lock "refresh_cache-ba8e47f7-e10d-439a-9a5a-714655a29dc0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.818060] env[63021]: DEBUG oslo_concurrency.lockutils [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Releasing lock "refresh_cache-7d98e2e4-33bc-4e6b-af6f-31c18f34b9af" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.818351] env[63021]: DEBUG nova.compute.manager [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 585.818527] env[63021]: DEBUG nova.compute.manager [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 585.818775] env[63021]: DEBUG nova.network.neutron [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 585.869505] env[63021]: DEBUG nova.network.neutron [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 585.924507] env[63021]: DEBUG nova.compute.manager [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 586.011906] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Task: {'id': task-1293560, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067122} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 586.012219] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 586.013347] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11418a90-34ea-485a-afae-420f71fc9bfd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.040304] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Reconfiguring VM instance instance-0000000b to attach disk [datastore2] 9d474bd0-e825-49f2-9715-14d85495718e/9d474bd0-e825-49f2-9715-14d85495718e.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 586.040899] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e54e58c-b9c6-4946-a3c9-779d648b110b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.063110] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Waiting for the task: (returnval){ [ 586.063110] env[63021]: value = "task-1293562" [ 586.063110] env[63021]: _type = "Task" [ 586.063110] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 586.071371] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Task: {'id': task-1293562, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 586.109162] env[63021]: DEBUG oslo_vmware.api [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293561, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.096152} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 586.109162] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 586.109162] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fadd6192-c1cb-4ea7-908d-7bc6d8844da2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.129615] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Reconfiguring VM instance instance-0000000a to attach disk [datastore2] 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd/5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 586.129914] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-73659149-e214-4c76-ada0-46a3ab3f87eb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.145254] env[63021]: DEBUG nova.scheduler.client.report [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 586.153640] env[63021]: DEBUG oslo_vmware.api [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 586.153640] env[63021]: value = "task-1293563" [ 586.153640] env[63021]: _type = "Task" [ 586.153640] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 586.162342] env[63021]: DEBUG oslo_vmware.api [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293563, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 586.303470] env[63021]: DEBUG nova.compute.manager [req-45b6bda6-1e4e-434c-8518-af646791d2f5 req-14507214-c856-4a52-b859-e09c259124da service nova] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Received event network-vif-deleted-40185534-7f2c-4524-98aa-8dad833d2c43 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 586.373142] env[63021]: DEBUG nova.network.neutron [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.456212] env[63021]: DEBUG oslo_concurrency.lockutils [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 586.580185] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Task: {'id': task-1293562, 'name': ReconfigVM_Task, 'duration_secs': 0.313592} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 586.580470] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Reconfigured VM instance instance-0000000b to attach disk [datastore2] 9d474bd0-e825-49f2-9715-14d85495718e/9d474bd0-e825-49f2-9715-14d85495718e.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 586.581117] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5cd7f0b6-ea81-416a-9143-993e86931b27 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.587035] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Waiting for the task: (returnval){ [ 586.587035] env[63021]: value = "task-1293565" [ 586.587035] env[63021]: _type = "Task" [ 586.587035] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 586.594874] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Task: {'id': task-1293565, 'name': Rename_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 586.651918] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.139s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.652604] env[63021]: ERROR nova.compute.manager [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f2cc2dc3-c7ef-4dcc-bbf3-c9154df16cc6, please check neutron logs for more information. [ 586.652604] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Traceback (most recent call last): [ 586.652604] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 586.652604] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] self.driver.spawn(context, instance, image_meta, [ 586.652604] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 586.652604] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] self._vmops.spawn(context, instance, image_meta, injected_files, [ 586.652604] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 586.652604] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] vm_ref = self.build_virtual_machine(instance, [ 586.652604] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 586.652604] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] vif_infos = vmwarevif.get_vif_info(self._session, [ 586.652604] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 586.652949] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] for vif in network_info: [ 586.652949] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 586.652949] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] return self._sync_wrapper(fn, *args, **kwargs) [ 586.652949] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 586.652949] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] self.wait() [ 586.652949] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 586.652949] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] self[:] = self._gt.wait() [ 586.652949] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 586.652949] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] return self._exit_event.wait() [ 586.652949] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 586.652949] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] result = hub.switch() [ 586.652949] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 586.652949] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] return self.greenlet.switch() [ 586.653360] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 586.653360] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] result = function(*args, **kwargs) [ 586.653360] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 586.653360] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] return func(*args, **kwargs) [ 586.653360] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 586.653360] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] raise e [ 586.653360] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 586.653360] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] nwinfo = self.network_api.allocate_for_instance( [ 586.653360] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 586.653360] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] created_port_ids = self._update_ports_for_instance( [ 586.653360] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 586.653360] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] with excutils.save_and_reraise_exception(): [ 586.653360] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.653710] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] self.force_reraise() [ 586.653710] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.653710] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] raise self.value [ 586.653710] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 586.653710] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] updated_port = self._update_port( [ 586.653710] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.653710] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] _ensure_no_port_binding_failure(port) [ 586.653710] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.653710] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] raise exception.PortBindingFailed(port_id=port['id']) [ 586.653710] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] nova.exception.PortBindingFailed: Binding failed for port f2cc2dc3-c7ef-4dcc-bbf3-c9154df16cc6, please check neutron logs for more information. [ 586.653710] env[63021]: ERROR nova.compute.manager [instance: ea199990-9638-4e94-87fe-bd4b7f020595] [ 586.654792] env[63021]: DEBUG nova.compute.utils [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Binding failed for port f2cc2dc3-c7ef-4dcc-bbf3-c9154df16cc6, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 586.656630] env[63021]: DEBUG oslo_concurrency.lockutils [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.322s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.660172] env[63021]: DEBUG nova.compute.manager [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Build of instance ea199990-9638-4e94-87fe-bd4b7f020595 was re-scheduled: Binding failed for port f2cc2dc3-c7ef-4dcc-bbf3-c9154df16cc6, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 586.660695] env[63021]: DEBUG nova.compute.manager [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 586.660906] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Acquiring lock "refresh_cache-ea199990-9638-4e94-87fe-bd4b7f020595" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.661245] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Acquired lock "refresh_cache-ea199990-9638-4e94-87fe-bd4b7f020595" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.661245] env[63021]: DEBUG nova.network.neutron [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 586.675703] env[63021]: DEBUG oslo_vmware.api [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293563, 'name': ReconfigVM_Task, 'duration_secs': 0.312748} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 586.676676] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Reconfigured VM instance instance-0000000a to attach disk [datastore2] 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd/5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 586.677323] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a7b7a1ec-6814-4888-b060-0fe5301682c7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.689379] env[63021]: DEBUG oslo_vmware.api [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 586.689379] env[63021]: value = "task-1293566" [ 586.689379] env[63021]: _type = "Task" [ 586.689379] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 586.699674] env[63021]: DEBUG oslo_vmware.api [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293566, 'name': Rename_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 586.823127] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Acquiring lock "39f51f90-a5a5-4479-881c-b5b23a3bf01a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 586.824099] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Lock "39f51f90-a5a5-4479-881c-b5b23a3bf01a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.876084] env[63021]: INFO nova.compute.manager [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] [instance: 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af] Took 1.06 seconds to deallocate network for instance. [ 587.099897] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Task: {'id': task-1293565, 'name': Rename_Task, 'duration_secs': 0.132294} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 587.101257] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 587.101958] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-539a8792-bccd-4409-a958-d1925fef39b2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.110877] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Waiting for the task: (returnval){ [ 587.110877] env[63021]: value = "task-1293567" [ 587.110877] env[63021]: _type = "Task" [ 587.110877] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 587.119010] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Task: {'id': task-1293567, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 587.198702] env[63021]: DEBUG oslo_vmware.api [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293566, 'name': Rename_Task, 'duration_secs': 0.141037} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 587.201835] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 587.202324] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-15080b1a-416f-43c5-bb01-4309425e36a9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.210722] env[63021]: DEBUG oslo_vmware.api [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 587.210722] env[63021]: value = "task-1293568" [ 587.210722] env[63021]: _type = "Task" [ 587.210722] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 587.226502] env[63021]: DEBUG oslo_vmware.api [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293568, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 587.232707] env[63021]: DEBUG nova.network.neutron [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 587.499379] env[63021]: DEBUG nova.network.neutron [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.623173] env[63021]: DEBUG oslo_vmware.api [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Task: {'id': task-1293567, 'name': PowerOnVM_Task, 'duration_secs': 0.498812} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 587.623467] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 587.623705] env[63021]: INFO nova.compute.manager [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Took 11.11 seconds to spawn the instance on the hypervisor. [ 587.623812] env[63021]: DEBUG nova.compute.manager [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 587.625740] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8f68348-c108-4d7a-806a-5ba3b1f18da7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.660522] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-738ad163-9b4a-478f-ab5a-e3eb3698d6b4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.668919] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7a18f2b-6f3c-4e6c-8e90-708e6b48b996 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.711794] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-611b4218-8514-4be4-a8f3-37d30507cc88 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.730019] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-161a5375-1df3-4be4-a7c0-8687a33baa34 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.735500] env[63021]: DEBUG oslo_vmware.api [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293568, 'name': PowerOnVM_Task, 'duration_secs': 0.505175} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 587.735886] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 587.736200] env[63021]: DEBUG nova.compute.manager [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 587.737398] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81478f83-19c6-4575-87a6-d29dfb2379e0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.749776] env[63021]: DEBUG nova.compute.provider_tree [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 587.941223] env[63021]: INFO nova.scheduler.client.report [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Deleted allocations for instance 7d98e2e4-33bc-4e6b-af6f-31c18f34b9af [ 588.004749] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Releasing lock "refresh_cache-ea199990-9638-4e94-87fe-bd4b7f020595" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 588.005357] env[63021]: DEBUG nova.compute.manager [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 588.005357] env[63021]: DEBUG nova.compute.manager [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 588.005357] env[63021]: DEBUG nova.network.neutron [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 588.073359] env[63021]: DEBUG nova.network.neutron [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 588.151909] env[63021]: INFO nova.compute.manager [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Took 31.61 seconds to build instance. [ 588.252653] env[63021]: DEBUG nova.scheduler.client.report [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 588.269165] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.457225] env[63021]: DEBUG oslo_concurrency.lockutils [None req-988f1e2c-f4a8-48a6-bd0e-17bea2a0d617 tempest-ServerDiagnosticsTest-1022122284 tempest-ServerDiagnosticsTest-1022122284-project-member] Lock "7d98e2e4-33bc-4e6b-af6f-31c18f34b9af" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.347s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.576152] env[63021]: DEBUG nova.network.neutron [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.639880] env[63021]: DEBUG nova.compute.manager [None req-fa4013ef-992f-4d86-aa85-80e0fbf7e7ee tempest-ServerDiagnosticsV248Test-740610966 tempest-ServerDiagnosticsV248Test-740610966-project-admin] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 588.642386] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d5b172a-5388-4460-9d29-77f26aa79095 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.653871] env[63021]: INFO nova.compute.manager [None req-fa4013ef-992f-4d86-aa85-80e0fbf7e7ee tempest-ServerDiagnosticsV248Test-740610966 tempest-ServerDiagnosticsV248Test-740610966-project-admin] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Retrieving diagnostics [ 588.653871] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7fcc7ed6-9904-4e75-89c9-ed288e92a00e tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Lock "9d474bd0-e825-49f2-9715-14d85495718e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.660s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.654593] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ddb5f8d-1d5a-4b83-8084-7c21fa811269 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.760260] env[63021]: DEBUG oslo_concurrency.lockutils [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.101s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.760260] env[63021]: ERROR nova.compute.manager [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 60d66191-46ca-4e31-8dda-90f2a0c92b59, please check neutron logs for more information. [ 588.760260] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Traceback (most recent call last): [ 588.760260] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 588.760260] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] self.driver.spawn(context, instance, image_meta, [ 588.760260] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 588.760260] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 588.760260] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 588.760260] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] vm_ref = self.build_virtual_machine(instance, [ 588.760595] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 588.760595] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] vif_infos = vmwarevif.get_vif_info(self._session, [ 588.760595] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 588.760595] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] for vif in network_info: [ 588.760595] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 588.760595] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] return self._sync_wrapper(fn, *args, **kwargs) [ 588.760595] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 588.760595] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] self.wait() [ 588.760595] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 588.760595] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] self[:] = self._gt.wait() [ 588.760595] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 588.760595] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] return self._exit_event.wait() [ 588.760595] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 588.760987] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] result = hub.switch() [ 588.760987] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 588.760987] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] return self.greenlet.switch() [ 588.760987] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 588.760987] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] result = function(*args, **kwargs) [ 588.760987] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 588.760987] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] return func(*args, **kwargs) [ 588.760987] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 588.760987] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] raise e [ 588.760987] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.760987] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] nwinfo = self.network_api.allocate_for_instance( [ 588.760987] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 588.760987] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] created_port_ids = self._update_ports_for_instance( [ 588.761341] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 588.761341] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] with excutils.save_and_reraise_exception(): [ 588.761341] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.761341] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] self.force_reraise() [ 588.761341] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.761341] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] raise self.value [ 588.761341] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 588.761341] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] updated_port = self._update_port( [ 588.761341] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.761341] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] _ensure_no_port_binding_failure(port) [ 588.761341] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.761341] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] raise exception.PortBindingFailed(port_id=port['id']) [ 588.761706] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] nova.exception.PortBindingFailed: Binding failed for port 60d66191-46ca-4e31-8dda-90f2a0c92b59, please check neutron logs for more information. [ 588.761706] env[63021]: ERROR nova.compute.manager [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] [ 588.761706] env[63021]: DEBUG nova.compute.utils [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Binding failed for port 60d66191-46ca-4e31-8dda-90f2a0c92b59, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 588.763295] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.410s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.765326] env[63021]: INFO nova.compute.claims [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 588.768355] env[63021]: DEBUG nova.compute.manager [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Build of instance 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9 was re-scheduled: Binding failed for port 60d66191-46ca-4e31-8dda-90f2a0c92b59, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 588.768913] env[63021]: DEBUG nova.compute.manager [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 588.769175] env[63021]: DEBUG oslo_concurrency.lockutils [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Acquiring lock "refresh_cache-2201799a-2df0-4fba-bcce-1fd8dfa1b6b9" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.769362] env[63021]: DEBUG oslo_concurrency.lockutils [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Acquired lock "refresh_cache-2201799a-2df0-4fba-bcce-1fd8dfa1b6b9" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.769554] env[63021]: DEBUG nova.network.neutron [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 588.961851] env[63021]: DEBUG nova.compute.manager [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 589.079638] env[63021]: INFO nova.compute.manager [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] [instance: ea199990-9638-4e94-87fe-bd4b7f020595] Took 1.07 seconds to deallocate network for instance. [ 589.163207] env[63021]: DEBUG nova.compute.manager [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 589.462331] env[63021]: DEBUG nova.network.neutron [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 589.506805] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.688763] env[63021]: DEBUG oslo_concurrency.lockutils [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.713052] env[63021]: DEBUG nova.network.neutron [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.118403] env[63021]: INFO nova.scheduler.client.report [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Deleted allocations for instance ea199990-9638-4e94-87fe-bd4b7f020595 [ 590.215434] env[63021]: DEBUG oslo_concurrency.lockutils [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Releasing lock "refresh_cache-2201799a-2df0-4fba-bcce-1fd8dfa1b6b9" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.215726] env[63021]: DEBUG nova.compute.manager [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 590.215830] env[63021]: DEBUG nova.compute.manager [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 590.215989] env[63021]: DEBUG nova.network.neutron [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 590.259925] env[63021]: DEBUG nova.network.neutron [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.286691] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d46ee8d-55e4-42ed-80f4-8bac34200025 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.298846] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6652ddee-b52e-42bc-8945-84bce6101b5d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.332219] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44f57db2-3712-4b95-a8f6-eb7b4cbb1866 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.342168] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81d2fd21-4839-42b8-8a88-475217a22144 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.357273] env[63021]: DEBUG nova.compute.provider_tree [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 590.639989] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ca6ffb15-f46b-4480-973d-c81169ff1119 tempest-ServerDiagnosticsNegativeTest-2086528500 tempest-ServerDiagnosticsNegativeTest-2086528500-project-member] Lock "ea199990-9638-4e94-87fe-bd4b7f020595" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.882s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.766653] env[63021]: DEBUG nova.network.neutron [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.861240] env[63021]: DEBUG nova.scheduler.client.report [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 591.143289] env[63021]: DEBUG nova.compute.manager [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 591.269628] env[63021]: INFO nova.compute.manager [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9] Took 1.05 seconds to deallocate network for instance. [ 591.373133] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.609s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.373133] env[63021]: DEBUG nova.compute.manager [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 591.383642] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.893s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.385311] env[63021]: INFO nova.compute.claims [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 591.500813] env[63021]: DEBUG oslo_concurrency.lockutils [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Acquiring lock "77fd1199-33b5-4040-a0ba-dd881da1224c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.502037] env[63021]: DEBUG oslo_concurrency.lockutils [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Lock "77fd1199-33b5-4040-a0ba-dd881da1224c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.519033] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Acquiring lock "5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.519033] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Lock "5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.519033] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Acquiring lock "5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.519033] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Lock "5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.519251] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Lock "5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.521960] env[63021]: INFO nova.compute.manager [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Terminating instance [ 591.524346] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Acquiring lock "refresh_cache-5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.524907] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Acquired lock "refresh_cache-5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.524907] env[63021]: DEBUG nova.network.neutron [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 591.677184] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.890775] env[63021]: DEBUG nova.compute.utils [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 591.898032] env[63021]: DEBUG nova.compute.manager [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 591.898032] env[63021]: DEBUG nova.network.neutron [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 592.047230] env[63021]: DEBUG nova.network.neutron [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.106926] env[63021]: DEBUG nova.network.neutron [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.196514] env[63021]: DEBUG nova.policy [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c35adcbb7506402fb1a457e803eb0b2a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '73afdfd26d4c4e4c81f183208cf9ab36', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 592.273621] env[63021]: DEBUG oslo_concurrency.lockutils [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Acquiring lock "4e6dcd18-4814-4cf1-966b-7d853858e650" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.273860] env[63021]: DEBUG oslo_concurrency.lockutils [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Lock "4e6dcd18-4814-4cf1-966b-7d853858e650" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.315954] env[63021]: INFO nova.scheduler.client.report [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Deleted allocations for instance 2201799a-2df0-4fba-bcce-1fd8dfa1b6b9 [ 592.404214] env[63021]: DEBUG nova.compute.manager [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 592.612938] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Releasing lock "refresh_cache-5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.612938] env[63021]: DEBUG nova.compute.manager [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 592.613924] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 592.616293] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1291872-40d8-4181-839e-23c34745b43f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.623910] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 592.624352] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f5cf5abc-8059-476a-845e-9a898aa3805a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.633118] env[63021]: DEBUG oslo_vmware.api [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 592.633118] env[63021]: value = "task-1293570" [ 592.633118] env[63021]: _type = "Task" [ 592.633118] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 592.647103] env[63021]: DEBUG oslo_vmware.api [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293570, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 592.824818] env[63021]: DEBUG oslo_concurrency.lockutils [None req-422a4fe4-b11b-4094-8dc6-ea6e1e278994 tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Lock "2201799a-2df0-4fba-bcce-1fd8dfa1b6b9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.231s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.854033] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2ac79fa-c939-40ab-9f65-79f0b4db45a6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.863633] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71a18bfd-a55d-47fa-8f03-bbc35a78bbaf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.905110] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25891484-e656-4763-b1dc-ebd4d0e17912 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.917620] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc30faa3-c46e-47b2-9f32-9214d20a19aa {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.936961] env[63021]: DEBUG nova.compute.provider_tree [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 593.076783] env[63021]: DEBUG nova.network.neutron [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Successfully created port: 9067a00a-e182-455c-88cb-1e0d7666dfe5 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 593.145645] env[63021]: DEBUG oslo_vmware.api [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293570, 'name': PowerOffVM_Task, 'duration_secs': 0.255763} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 593.145988] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 593.146404] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 593.146851] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-21e5eb0c-7674-4035-97e7-c218e784fb82 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.173051] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 593.173292] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Deleting contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 593.173416] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Deleting the datastore file [datastore2] 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 593.173667] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d82d92d6-d000-4b8e-bde4-ef2810cdacd3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.181176] env[63021]: DEBUG oslo_vmware.api [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 593.181176] env[63021]: value = "task-1293572" [ 593.181176] env[63021]: _type = "Task" [ 593.181176] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 593.190882] env[63021]: DEBUG oslo_vmware.api [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293572, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 593.328257] env[63021]: DEBUG nova.compute.manager [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 593.418339] env[63021]: DEBUG nova.compute.manager [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 593.425734] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "52b854c4-ccc8-4e38-8100-31ab06959939" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.426103] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "52b854c4-ccc8-4e38-8100-31ab06959939" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.439987] env[63021]: DEBUG nova.scheduler.client.report [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 593.458614] env[63021]: DEBUG nova.virt.hardware [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:44:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1067201452',id=28,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-1757899554',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 593.458915] env[63021]: DEBUG nova.virt.hardware [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 593.459954] env[63021]: DEBUG nova.virt.hardware [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 593.459954] env[63021]: DEBUG nova.virt.hardware [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 593.459954] env[63021]: DEBUG nova.virt.hardware [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 593.460295] env[63021]: DEBUG nova.virt.hardware [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 593.460600] env[63021]: DEBUG nova.virt.hardware [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 593.460894] env[63021]: DEBUG nova.virt.hardware [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 593.461474] env[63021]: DEBUG nova.virt.hardware [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 593.461688] env[63021]: DEBUG nova.virt.hardware [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 593.461901] env[63021]: DEBUG nova.virt.hardware [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 593.463652] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-144a0595-9ded-4027-9558-153daf5f7cc9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.474417] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbb3b024-7cce-49a1-b817-ed38ad2b75dc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.693303] env[63021]: DEBUG oslo_vmware.api [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293572, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.112964} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 593.693623] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 593.693808] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Deleted contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 593.693984] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 593.694393] env[63021]: INFO nova.compute.manager [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Took 1.08 seconds to destroy the instance on the hypervisor. [ 593.694662] env[63021]: DEBUG oslo.service.loopingcall [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 593.694853] env[63021]: DEBUG nova.compute.manager [-] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 593.694950] env[63021]: DEBUG nova.network.neutron [-] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 593.725927] env[63021]: DEBUG nova.network.neutron [-] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 593.861775] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.948564] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.562s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.948564] env[63021]: DEBUG nova.compute.manager [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 593.953480] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.252s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.953480] env[63021]: INFO nova.compute.claims [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 594.228018] env[63021]: DEBUG nova.network.neutron [-] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.301577] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Acquiring lock "aef64522-2dbb-4bc5-84b2-a0bf8a108574" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.303050] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Lock "aef64522-2dbb-4bc5-84b2-a0bf8a108574" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.467014] env[63021]: DEBUG nova.compute.utils [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 594.468807] env[63021]: DEBUG nova.compute.manager [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 594.468985] env[63021]: DEBUG nova.network.neutron [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 594.482340] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Acquiring lock "c6ed1f3b-e80e-4545-8338-8e3aead5ba16" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.485019] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Lock "c6ed1f3b-e80e-4545-8338-8e3aead5ba16" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.515346] env[63021]: DEBUG nova.policy [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '01b2489aab484c0091e9d3c3c5b8e77e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6ac184b54b43453b8ad886e2979edd80', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 594.730945] env[63021]: INFO nova.compute.manager [-] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Took 1.04 seconds to deallocate network for instance. [ 594.984870] env[63021]: DEBUG nova.compute.manager [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 595.058985] env[63021]: DEBUG nova.network.neutron [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Successfully created port: 95e81d9c-71a8-4616-a6a9-1cc0eeef6164 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 595.126208] env[63021]: DEBUG nova.compute.manager [req-58157f87-af11-4489-8ec4-bb180d43be9a req-61f13779-bee8-4010-8ab0-cfead89b3a8f service nova] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Received event network-changed-9067a00a-e182-455c-88cb-1e0d7666dfe5 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 595.126394] env[63021]: DEBUG nova.compute.manager [req-58157f87-af11-4489-8ec4-bb180d43be9a req-61f13779-bee8-4010-8ab0-cfead89b3a8f service nova] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Refreshing instance network info cache due to event network-changed-9067a00a-e182-455c-88cb-1e0d7666dfe5. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 595.126604] env[63021]: DEBUG oslo_concurrency.lockutils [req-58157f87-af11-4489-8ec4-bb180d43be9a req-61f13779-bee8-4010-8ab0-cfead89b3a8f service nova] Acquiring lock "refresh_cache-082e6e14-c5f7-4ca4-b739-9d997a77b8f9" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.126741] env[63021]: DEBUG oslo_concurrency.lockutils [req-58157f87-af11-4489-8ec4-bb180d43be9a req-61f13779-bee8-4010-8ab0-cfead89b3a8f service nova] Acquired lock "refresh_cache-082e6e14-c5f7-4ca4-b739-9d997a77b8f9" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.126896] env[63021]: DEBUG nova.network.neutron [req-58157f87-af11-4489-8ec4-bb180d43be9a req-61f13779-bee8-4010-8ab0-cfead89b3a8f service nova] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Refreshing network info cache for port 9067a00a-e182-455c-88cb-1e0d7666dfe5 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 595.237677] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.453816] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-404c26ef-e005-4acb-9713-844da54c0d01 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.463039] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7743b4e4-6768-470d-9022-8d5bf40f0547 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.507163] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28bdeb41-d0c7-4e60-aa81-3e8dbbdcf5d7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.516693] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9efedafd-be8b-4ec2-90f2-ac2fa9d652b8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.531764] env[63021]: DEBUG nova.compute.provider_tree [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 595.634372] env[63021]: ERROR nova.compute.manager [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9067a00a-e182-455c-88cb-1e0d7666dfe5, please check neutron logs for more information. [ 595.634372] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 595.634372] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 595.634372] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 595.634372] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 595.634372] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 595.634372] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 595.634372] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 595.634372] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.634372] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 595.634372] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.634372] env[63021]: ERROR nova.compute.manager raise self.value [ 595.634372] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 595.634372] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 595.634372] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.634372] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 595.634992] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.634992] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 595.634992] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9067a00a-e182-455c-88cb-1e0d7666dfe5, please check neutron logs for more information. [ 595.634992] env[63021]: ERROR nova.compute.manager [ 595.634992] env[63021]: Traceback (most recent call last): [ 595.634992] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 595.634992] env[63021]: listener.cb(fileno) [ 595.634992] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 595.634992] env[63021]: result = function(*args, **kwargs) [ 595.634992] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 595.634992] env[63021]: return func(*args, **kwargs) [ 595.634992] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 595.634992] env[63021]: raise e [ 595.634992] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 595.634992] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 595.634992] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 595.634992] env[63021]: created_port_ids = self._update_ports_for_instance( [ 595.634992] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 595.634992] env[63021]: with excutils.save_and_reraise_exception(): [ 595.634992] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.634992] env[63021]: self.force_reraise() [ 595.634992] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.634992] env[63021]: raise self.value [ 595.634992] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 595.634992] env[63021]: updated_port = self._update_port( [ 595.634992] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.634992] env[63021]: _ensure_no_port_binding_failure(port) [ 595.634992] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.634992] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 595.637953] env[63021]: nova.exception.PortBindingFailed: Binding failed for port 9067a00a-e182-455c-88cb-1e0d7666dfe5, please check neutron logs for more information. [ 595.637953] env[63021]: Removing descriptor: 17 [ 595.637953] env[63021]: ERROR nova.compute.manager [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9067a00a-e182-455c-88cb-1e0d7666dfe5, please check neutron logs for more information. [ 595.637953] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Traceback (most recent call last): [ 595.637953] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 595.637953] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] yield resources [ 595.637953] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 595.637953] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] self.driver.spawn(context, instance, image_meta, [ 595.637953] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 595.637953] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 595.637953] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 595.637953] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] vm_ref = self.build_virtual_machine(instance, [ 595.638453] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 595.638453] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] vif_infos = vmwarevif.get_vif_info(self._session, [ 595.638453] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 595.638453] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] for vif in network_info: [ 595.638453] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 595.638453] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] return self._sync_wrapper(fn, *args, **kwargs) [ 595.638453] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 595.638453] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] self.wait() [ 595.638453] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 595.638453] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] self[:] = self._gt.wait() [ 595.638453] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 595.638453] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] return self._exit_event.wait() [ 595.638453] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 595.638856] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] result = hub.switch() [ 595.638856] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 595.638856] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] return self.greenlet.switch() [ 595.638856] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 595.638856] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] result = function(*args, **kwargs) [ 595.638856] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 595.638856] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] return func(*args, **kwargs) [ 595.638856] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 595.638856] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] raise e [ 595.638856] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 595.638856] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] nwinfo = self.network_api.allocate_for_instance( [ 595.638856] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 595.638856] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] created_port_ids = self._update_ports_for_instance( [ 595.639292] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 595.639292] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] with excutils.save_and_reraise_exception(): [ 595.639292] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.639292] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] self.force_reraise() [ 595.639292] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.639292] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] raise self.value [ 595.639292] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 595.639292] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] updated_port = self._update_port( [ 595.639292] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.639292] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] _ensure_no_port_binding_failure(port) [ 595.639292] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.639292] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] raise exception.PortBindingFailed(port_id=port['id']) [ 595.639658] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] nova.exception.PortBindingFailed: Binding failed for port 9067a00a-e182-455c-88cb-1e0d7666dfe5, please check neutron logs for more information. [ 595.639658] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] [ 595.639658] env[63021]: INFO nova.compute.manager [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Terminating instance [ 595.639658] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Acquiring lock "refresh_cache-082e6e14-c5f7-4ca4-b739-9d997a77b8f9" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.660320] env[63021]: DEBUG nova.network.neutron [req-58157f87-af11-4489-8ec4-bb180d43be9a req-61f13779-bee8-4010-8ab0-cfead89b3a8f service nova] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 595.937377] env[63021]: DEBUG nova.network.neutron [req-58157f87-af11-4489-8ec4-bb180d43be9a req-61f13779-bee8-4010-8ab0-cfead89b3a8f service nova] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.006736] env[63021]: DEBUG nova.compute.manager [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 596.037195] env[63021]: DEBUG nova.scheduler.client.report [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 596.063107] env[63021]: DEBUG nova.virt.hardware [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 596.063107] env[63021]: DEBUG nova.virt.hardware [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 596.063107] env[63021]: DEBUG nova.virt.hardware [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 596.063781] env[63021]: DEBUG nova.virt.hardware [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 596.064101] env[63021]: DEBUG nova.virt.hardware [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 596.064420] env[63021]: DEBUG nova.virt.hardware [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 596.066508] env[63021]: DEBUG nova.virt.hardware [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 596.066508] env[63021]: DEBUG nova.virt.hardware [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 596.066508] env[63021]: DEBUG nova.virt.hardware [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 596.066508] env[63021]: DEBUG nova.virt.hardware [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 596.066508] env[63021]: DEBUG nova.virt.hardware [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 596.067459] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af14b439-337d-4ac1-b12d-092c9d61ded8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.080515] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca8646be-ec4c-4dab-a715-badc0cb733bf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.439168] env[63021]: DEBUG oslo_concurrency.lockutils [req-58157f87-af11-4489-8ec4-bb180d43be9a req-61f13779-bee8-4010-8ab0-cfead89b3a8f service nova] Releasing lock "refresh_cache-082e6e14-c5f7-4ca4-b739-9d997a77b8f9" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.439601] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Acquired lock "refresh_cache-082e6e14-c5f7-4ca4-b739-9d997a77b8f9" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.439779] env[63021]: DEBUG nova.network.neutron [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 596.546055] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.595s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 596.546274] env[63021]: DEBUG nova.compute.manager [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 596.548926] env[63021]: DEBUG oslo_concurrency.lockutils [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.267s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.550344] env[63021]: INFO nova.compute.claims [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 596.975918] env[63021]: DEBUG nova.network.neutron [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 597.058244] env[63021]: DEBUG nova.compute.utils [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 597.061436] env[63021]: DEBUG nova.compute.manager [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 597.061606] env[63021]: DEBUG nova.network.neutron [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 597.131250] env[63021]: DEBUG nova.policy [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '93af6011875d49a9a99c660d4b263a95', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b6bfffa3ff494e799c401542a886a945', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 597.197539] env[63021]: DEBUG nova.network.neutron [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.214166] env[63021]: ERROR nova.compute.manager [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 95e81d9c-71a8-4616-a6a9-1cc0eeef6164, please check neutron logs for more information. [ 597.214166] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 597.214166] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.214166] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 597.214166] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 597.214166] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 597.214166] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 597.214166] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 597.214166] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.214166] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 597.214166] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.214166] env[63021]: ERROR nova.compute.manager raise self.value [ 597.214166] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 597.214166] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 597.214166] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.214166] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 597.214707] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.214707] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 597.214707] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 95e81d9c-71a8-4616-a6a9-1cc0eeef6164, please check neutron logs for more information. [ 597.214707] env[63021]: ERROR nova.compute.manager [ 597.214707] env[63021]: Traceback (most recent call last): [ 597.214707] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 597.214707] env[63021]: listener.cb(fileno) [ 597.214707] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.214707] env[63021]: result = function(*args, **kwargs) [ 597.214707] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 597.214707] env[63021]: return func(*args, **kwargs) [ 597.214707] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 597.214707] env[63021]: raise e [ 597.214707] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.214707] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 597.214707] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 597.214707] env[63021]: created_port_ids = self._update_ports_for_instance( [ 597.214707] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 597.214707] env[63021]: with excutils.save_and_reraise_exception(): [ 597.214707] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.214707] env[63021]: self.force_reraise() [ 597.214707] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.214707] env[63021]: raise self.value [ 597.214707] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 597.214707] env[63021]: updated_port = self._update_port( [ 597.214707] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.214707] env[63021]: _ensure_no_port_binding_failure(port) [ 597.214707] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.214707] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 597.215678] env[63021]: nova.exception.PortBindingFailed: Binding failed for port 95e81d9c-71a8-4616-a6a9-1cc0eeef6164, please check neutron logs for more information. [ 597.215678] env[63021]: Removing descriptor: 15 [ 597.215678] env[63021]: ERROR nova.compute.manager [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 95e81d9c-71a8-4616-a6a9-1cc0eeef6164, please check neutron logs for more information. [ 597.215678] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Traceback (most recent call last): [ 597.215678] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 597.215678] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] yield resources [ 597.215678] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 597.215678] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] self.driver.spawn(context, instance, image_meta, [ 597.215678] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 597.215678] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 597.215678] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 597.215678] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] vm_ref = self.build_virtual_machine(instance, [ 597.216057] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 597.216057] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] vif_infos = vmwarevif.get_vif_info(self._session, [ 597.216057] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 597.216057] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] for vif in network_info: [ 597.216057] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 597.216057] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] return self._sync_wrapper(fn, *args, **kwargs) [ 597.216057] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 597.216057] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] self.wait() [ 597.216057] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 597.216057] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] self[:] = self._gt.wait() [ 597.216057] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 597.216057] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] return self._exit_event.wait() [ 597.216057] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 597.216484] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] result = hub.switch() [ 597.216484] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 597.216484] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] return self.greenlet.switch() [ 597.216484] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.216484] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] result = function(*args, **kwargs) [ 597.216484] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 597.216484] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] return func(*args, **kwargs) [ 597.216484] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 597.216484] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] raise e [ 597.216484] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.216484] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] nwinfo = self.network_api.allocate_for_instance( [ 597.216484] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 597.216484] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] created_port_ids = self._update_ports_for_instance( [ 597.216913] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 597.216913] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] with excutils.save_and_reraise_exception(): [ 597.216913] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.216913] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] self.force_reraise() [ 597.216913] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.216913] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] raise self.value [ 597.216913] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 597.216913] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] updated_port = self._update_port( [ 597.216913] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.216913] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] _ensure_no_port_binding_failure(port) [ 597.216913] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.216913] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] raise exception.PortBindingFailed(port_id=port['id']) [ 597.217325] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] nova.exception.PortBindingFailed: Binding failed for port 95e81d9c-71a8-4616-a6a9-1cc0eeef6164, please check neutron logs for more information. [ 597.217325] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] [ 597.217325] env[63021]: INFO nova.compute.manager [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Terminating instance [ 597.220522] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Acquiring lock "refresh_cache-9f64a5ad-7852-4772-9b8c-2ec1537204f3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.220522] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Acquired lock "refresh_cache-9f64a5ad-7852-4772-9b8c-2ec1537204f3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.220522] env[63021]: DEBUG nova.network.neutron [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 597.343123] env[63021]: DEBUG nova.compute.manager [req-d9ffbb3b-3b3e-4d83-916e-58bec3f741f4 req-9aee6c66-2c12-42b1-865e-11031c029571 service nova] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Received event network-changed-95e81d9c-71a8-4616-a6a9-1cc0eeef6164 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 597.343123] env[63021]: DEBUG nova.compute.manager [req-d9ffbb3b-3b3e-4d83-916e-58bec3f741f4 req-9aee6c66-2c12-42b1-865e-11031c029571 service nova] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Refreshing instance network info cache due to event network-changed-95e81d9c-71a8-4616-a6a9-1cc0eeef6164. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 597.343500] env[63021]: DEBUG oslo_concurrency.lockutils [req-d9ffbb3b-3b3e-4d83-916e-58bec3f741f4 req-9aee6c66-2c12-42b1-865e-11031c029571 service nova] Acquiring lock "refresh_cache-9f64a5ad-7852-4772-9b8c-2ec1537204f3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.522984] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Acquiring lock "d36a9ed2-6849-43fc-ab88-77468bde337f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.523220] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Lock "d36a9ed2-6849-43fc-ab88-77468bde337f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.562132] env[63021]: DEBUG nova.compute.manager [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 597.704183] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Releasing lock "refresh_cache-082e6e14-c5f7-4ca4-b739-9d997a77b8f9" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 597.705623] env[63021]: DEBUG nova.compute.manager [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 597.705623] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 597.705623] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6c21d51e-89f4-4d71-8d16-6405c8e252b6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.720189] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35330290-c536-4f6b-ab12-50067d530328 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.741697] env[63021]: DEBUG nova.network.neutron [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Successfully created port: 663a49a6-fb86-4199-b1f0-fbf8f32e806c {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 597.756683] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 082e6e14-c5f7-4ca4-b739-9d997a77b8f9 could not be found. [ 597.756919] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 597.757191] env[63021]: INFO nova.compute.manager [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Took 0.05 seconds to destroy the instance on the hypervisor. [ 597.757482] env[63021]: DEBUG oslo.service.loopingcall [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 597.757712] env[63021]: DEBUG nova.compute.manager [-] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 597.757790] env[63021]: DEBUG nova.network.neutron [-] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 597.855314] env[63021]: DEBUG nova.network.neutron [-] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 597.860255] env[63021]: DEBUG nova.network.neutron [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 597.863119] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquiring lock "476e1e73-7050-4e53-bff8-d5d9f0982f1a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.863475] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lock "476e1e73-7050-4e53-bff8-d5d9f0982f1a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.929653] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquiring lock "28d70e1d-da2a-4c9c-b297-86d026ce42c3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.929925] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lock "28d70e1d-da2a-4c9c-b297-86d026ce42c3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.000668] env[63021]: DEBUG nova.network.neutron [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.025065] env[63021]: DEBUG nova.compute.manager [req-8c11680c-8c79-4279-ac24-a40ac64ec904 req-d2bb8caf-a993-4890-8820-1c21f79e8d35 service nova] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Received event network-vif-deleted-9067a00a-e182-455c-88cb-1e0d7666dfe5 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 598.081234] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5962b1ea-36c3-4ad4-8abd-af0cc3501def {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.087934] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-555038b7-6e76-4947-8ed2-05c538f3da4f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.125083] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a298a4d4-46b8-49b0-836b-2506b04c0ba1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.133509] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94721fcb-299a-4864-8824-5023176d819a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.152333] env[63021]: DEBUG nova.compute.provider_tree [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 598.365358] env[63021]: DEBUG nova.network.neutron [-] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.453197] env[63021]: DEBUG oslo_concurrency.lockutils [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Acquiring lock "03aec3ed-f338-41b5-a396-e3d693edbb8c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.453473] env[63021]: DEBUG oslo_concurrency.lockutils [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Lock "03aec3ed-f338-41b5-a396-e3d693edbb8c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.503497] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Releasing lock "refresh_cache-9f64a5ad-7852-4772-9b8c-2ec1537204f3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.503962] env[63021]: DEBUG nova.compute.manager [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 598.504126] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 598.504435] env[63021]: DEBUG oslo_concurrency.lockutils [req-d9ffbb3b-3b3e-4d83-916e-58bec3f741f4 req-9aee6c66-2c12-42b1-865e-11031c029571 service nova] Acquired lock "refresh_cache-9f64a5ad-7852-4772-9b8c-2ec1537204f3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.506460] env[63021]: DEBUG nova.network.neutron [req-d9ffbb3b-3b3e-4d83-916e-58bec3f741f4 req-9aee6c66-2c12-42b1-865e-11031c029571 service nova] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Refreshing network info cache for port 95e81d9c-71a8-4616-a6a9-1cc0eeef6164 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 598.506460] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6b0188d4-f1d0-4293-b8cf-268a5b0c4844 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.525541] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eccad71e-29b1-46bd-bb94-31192fb27294 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.552951] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9f64a5ad-7852-4772-9b8c-2ec1537204f3 could not be found. [ 598.553197] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 598.553403] env[63021]: INFO nova.compute.manager [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Took 0.05 seconds to destroy the instance on the hypervisor. [ 598.553647] env[63021]: DEBUG oslo.service.loopingcall [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 598.553864] env[63021]: DEBUG nova.compute.manager [-] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 598.553957] env[63021]: DEBUG nova.network.neutron [-] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 598.577414] env[63021]: DEBUG nova.network.neutron [-] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.582021] env[63021]: DEBUG nova.compute.manager [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 598.604134] env[63021]: DEBUG nova.virt.hardware [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 598.604434] env[63021]: DEBUG nova.virt.hardware [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 598.604704] env[63021]: DEBUG nova.virt.hardware [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 598.604904] env[63021]: DEBUG nova.virt.hardware [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 598.605135] env[63021]: DEBUG nova.virt.hardware [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 598.605291] env[63021]: DEBUG nova.virt.hardware [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 598.605499] env[63021]: DEBUG nova.virt.hardware [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 598.605660] env[63021]: DEBUG nova.virt.hardware [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 598.606106] env[63021]: DEBUG nova.virt.hardware [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 598.606291] env[63021]: DEBUG nova.virt.hardware [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 598.606457] env[63021]: DEBUG nova.virt.hardware [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 598.608173] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79eda747-a54e-47a3-b6bc-f8343498ff76 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.619901] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd029aef-c8ab-4934-9fcc-d540f9bb7d55 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.652210] env[63021]: DEBUG nova.scheduler.client.report [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 598.868408] env[63021]: INFO nova.compute.manager [-] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Took 1.11 seconds to deallocate network for instance. [ 598.870953] env[63021]: DEBUG nova.compute.claims [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 598.871186] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.029815] env[63021]: DEBUG nova.network.neutron [req-d9ffbb3b-3b3e-4d83-916e-58bec3f741f4 req-9aee6c66-2c12-42b1-865e-11031c029571 service nova] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 599.079807] env[63021]: ERROR nova.compute.manager [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 663a49a6-fb86-4199-b1f0-fbf8f32e806c, please check neutron logs for more information. [ 599.079807] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 599.079807] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.079807] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 599.079807] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 599.079807] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 599.079807] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 599.079807] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 599.079807] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.079807] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 599.079807] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.079807] env[63021]: ERROR nova.compute.manager raise self.value [ 599.079807] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 599.079807] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 599.079807] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.079807] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 599.080340] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.080340] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 599.080340] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 663a49a6-fb86-4199-b1f0-fbf8f32e806c, please check neutron logs for more information. [ 599.080340] env[63021]: ERROR nova.compute.manager [ 599.080340] env[63021]: Traceback (most recent call last): [ 599.080340] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 599.080340] env[63021]: listener.cb(fileno) [ 599.080340] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 599.080340] env[63021]: result = function(*args, **kwargs) [ 599.080340] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 599.080340] env[63021]: return func(*args, **kwargs) [ 599.080340] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 599.080340] env[63021]: raise e [ 599.080340] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.080340] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 599.080340] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 599.080340] env[63021]: created_port_ids = self._update_ports_for_instance( [ 599.080340] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 599.080340] env[63021]: with excutils.save_and_reraise_exception(): [ 599.080340] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.080340] env[63021]: self.force_reraise() [ 599.080340] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.080340] env[63021]: raise self.value [ 599.080340] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 599.080340] env[63021]: updated_port = self._update_port( [ 599.080340] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.080340] env[63021]: _ensure_no_port_binding_failure(port) [ 599.080340] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.080340] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 599.081200] env[63021]: nova.exception.PortBindingFailed: Binding failed for port 663a49a6-fb86-4199-b1f0-fbf8f32e806c, please check neutron logs for more information. [ 599.081200] env[63021]: Removing descriptor: 17 [ 599.081200] env[63021]: ERROR nova.compute.manager [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 663a49a6-fb86-4199-b1f0-fbf8f32e806c, please check neutron logs for more information. [ 599.081200] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Traceback (most recent call last): [ 599.081200] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 599.081200] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] yield resources [ 599.081200] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 599.081200] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] self.driver.spawn(context, instance, image_meta, [ 599.081200] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 599.081200] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] self._vmops.spawn(context, instance, image_meta, injected_files, [ 599.081200] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 599.081200] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] vm_ref = self.build_virtual_machine(instance, [ 599.081716] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 599.081716] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] vif_infos = vmwarevif.get_vif_info(self._session, [ 599.081716] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 599.081716] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] for vif in network_info: [ 599.081716] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 599.081716] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] return self._sync_wrapper(fn, *args, **kwargs) [ 599.081716] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 599.081716] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] self.wait() [ 599.081716] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 599.081716] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] self[:] = self._gt.wait() [ 599.081716] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 599.081716] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] return self._exit_event.wait() [ 599.081716] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 599.082168] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] result = hub.switch() [ 599.082168] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 599.082168] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] return self.greenlet.switch() [ 599.082168] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 599.082168] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] result = function(*args, **kwargs) [ 599.082168] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 599.082168] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] return func(*args, **kwargs) [ 599.082168] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 599.082168] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] raise e [ 599.082168] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.082168] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] nwinfo = self.network_api.allocate_for_instance( [ 599.082168] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 599.082168] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] created_port_ids = self._update_ports_for_instance( [ 599.082563] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 599.082563] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] with excutils.save_and_reraise_exception(): [ 599.082563] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.082563] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] self.force_reraise() [ 599.082563] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.082563] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] raise self.value [ 599.082563] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 599.082563] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] updated_port = self._update_port( [ 599.082563] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.082563] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] _ensure_no_port_binding_failure(port) [ 599.082563] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.082563] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] raise exception.PortBindingFailed(port_id=port['id']) [ 599.082903] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] nova.exception.PortBindingFailed: Binding failed for port 663a49a6-fb86-4199-b1f0-fbf8f32e806c, please check neutron logs for more information. [ 599.082903] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] [ 599.082903] env[63021]: INFO nova.compute.manager [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Terminating instance [ 599.082903] env[63021]: DEBUG nova.network.neutron [-] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.083890] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Acquiring lock "refresh_cache-1fa1c7c7-b9fe-4e60-8306-3c263c517c28" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 599.084057] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Acquired lock "refresh_cache-1fa1c7c7-b9fe-4e60-8306-3c263c517c28" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.084241] env[63021]: DEBUG nova.network.neutron [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 599.159613] env[63021]: DEBUG oslo_concurrency.lockutils [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.611s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.160157] env[63021]: DEBUG nova.compute.manager [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 599.163295] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.643s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.169506] env[63021]: DEBUG nova.network.neutron [req-d9ffbb3b-3b3e-4d83-916e-58bec3f741f4 req-9aee6c66-2c12-42b1-865e-11031c029571 service nova] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.293028] env[63021]: DEBUG nova.compute.manager [None req-1490f968-d21c-46ef-ac48-a32657d2a4b5 tempest-ServerDiagnosticsV248Test-740610966 tempest-ServerDiagnosticsV248Test-740610966-project-admin] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 599.295592] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7be0bf41-c3bf-4af6-a6ba-8cd546b07463 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.302880] env[63021]: INFO nova.compute.manager [None req-1490f968-d21c-46ef-ac48-a32657d2a4b5 tempest-ServerDiagnosticsV248Test-740610966 tempest-ServerDiagnosticsV248Test-740610966-project-admin] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Retrieving diagnostics [ 599.303889] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-886212b2-8ef7-4533-a8b4-42e754aecc0a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.585891] env[63021]: INFO nova.compute.manager [-] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Took 1.03 seconds to deallocate network for instance. [ 599.590411] env[63021]: DEBUG nova.compute.claims [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 599.590621] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.607181] env[63021]: DEBUG nova.network.neutron [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 599.667241] env[63021]: DEBUG nova.compute.utils [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 599.669278] env[63021]: DEBUG nova.compute.manager [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 599.669278] env[63021]: DEBUG nova.network.neutron [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 599.674764] env[63021]: DEBUG oslo_concurrency.lockutils [req-d9ffbb3b-3b3e-4d83-916e-58bec3f741f4 req-9aee6c66-2c12-42b1-865e-11031c029571 service nova] Releasing lock "refresh_cache-9f64a5ad-7852-4772-9b8c-2ec1537204f3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.683284] env[63021]: DEBUG nova.network.neutron [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.737533] env[63021]: DEBUG nova.policy [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dfd31b7d3d0040aeb930085215c84f9f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9c5a44a56a2944d4ba9fac8f2a360fc2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 599.895538] env[63021]: DEBUG nova.compute.manager [req-07637f2d-8270-4d81-b734-3027d490aef8 req-199a3ed5-c499-4426-9595-e119976ef676 service nova] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Received event network-vif-deleted-95e81d9c-71a8-4616-a6a9-1cc0eeef6164 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 600.174415] env[63021]: DEBUG nova.compute.manager [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 600.179678] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2e6d0dd-a5ac-43d8-85b6-c4103136cf53 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.191326] env[63021]: DEBUG nova.network.neutron [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Successfully created port: 5522f00b-405a-4de6-9f40-35ad30faab5c {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 600.192777] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Releasing lock "refresh_cache-1fa1c7c7-b9fe-4e60-8306-3c263c517c28" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 600.193110] env[63021]: DEBUG nova.compute.manager [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 600.193293] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 600.193612] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-666bbd21-3e9d-40f4-b630-e59e744e2458 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.196575] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23b8cfc4-4723-4cf1-b185-63643ed49eab {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.231143] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11ed1fbe-4a5a-4840-94d1-c70159c02af7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.242917] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17abe84a-e350-4a23-95d4-eb640f084297 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.251074] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1261c41c-e027-4fad-ba6b-a134780b7734 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.261575] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1fa1c7c7-b9fe-4e60-8306-3c263c517c28 could not be found. [ 600.261803] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 600.262018] env[63021]: INFO nova.compute.manager [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Took 0.07 seconds to destroy the instance on the hypervisor. [ 600.262264] env[63021]: DEBUG oslo.service.loopingcall [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 600.262901] env[63021]: DEBUG nova.compute.manager [-] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 600.263036] env[63021]: DEBUG nova.network.neutron [-] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 600.273143] env[63021]: DEBUG nova.compute.provider_tree [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 600.291021] env[63021]: DEBUG nova.network.neutron [-] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 600.611153] env[63021]: DEBUG oslo_concurrency.lockutils [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Acquiring lock "239d0087-462d-445a-a131-a23d4d15505f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.611515] env[63021]: DEBUG oslo_concurrency.lockutils [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Lock "239d0087-462d-445a-a131-a23d4d15505f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.678803] env[63021]: DEBUG nova.compute.manager [req-a1ec9292-2bbe-4727-ad80-a140e34de000 req-d2ef1d10-27a1-491f-858c-de93eeaa97cd service nova] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Received event network-changed-663a49a6-fb86-4199-b1f0-fbf8f32e806c {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 600.679257] env[63021]: DEBUG nova.compute.manager [req-a1ec9292-2bbe-4727-ad80-a140e34de000 req-d2ef1d10-27a1-491f-858c-de93eeaa97cd service nova] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Refreshing instance network info cache due to event network-changed-663a49a6-fb86-4199-b1f0-fbf8f32e806c. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 600.679527] env[63021]: DEBUG oslo_concurrency.lockutils [req-a1ec9292-2bbe-4727-ad80-a140e34de000 req-d2ef1d10-27a1-491f-858c-de93eeaa97cd service nova] Acquiring lock "refresh_cache-1fa1c7c7-b9fe-4e60-8306-3c263c517c28" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 600.679631] env[63021]: DEBUG oslo_concurrency.lockutils [req-a1ec9292-2bbe-4727-ad80-a140e34de000 req-d2ef1d10-27a1-491f-858c-de93eeaa97cd service nova] Acquired lock "refresh_cache-1fa1c7c7-b9fe-4e60-8306-3c263c517c28" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.679788] env[63021]: DEBUG nova.network.neutron [req-a1ec9292-2bbe-4727-ad80-a140e34de000 req-d2ef1d10-27a1-491f-858c-de93eeaa97cd service nova] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Refreshing network info cache for port 663a49a6-fb86-4199-b1f0-fbf8f32e806c {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 600.775961] env[63021]: DEBUG nova.scheduler.client.report [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 600.792772] env[63021]: DEBUG nova.network.neutron [-] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.073552] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Acquiring lock "9d474bd0-e825-49f2-9715-14d85495718e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.073552] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Lock "9d474bd0-e825-49f2-9715-14d85495718e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 601.073552] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Acquiring lock "9d474bd0-e825-49f2-9715-14d85495718e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.073552] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Lock "9d474bd0-e825-49f2-9715-14d85495718e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 601.073801] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Lock "9d474bd0-e825-49f2-9715-14d85495718e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 601.075799] env[63021]: INFO nova.compute.manager [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Terminating instance [ 601.080999] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Acquiring lock "refresh_cache-9d474bd0-e825-49f2-9715-14d85495718e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.080999] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Acquired lock "refresh_cache-9d474bd0-e825-49f2-9715-14d85495718e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.081196] env[63021]: DEBUG nova.network.neutron [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 601.193721] env[63021]: DEBUG nova.compute.manager [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 601.228550] env[63021]: DEBUG nova.virt.hardware [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 601.228888] env[63021]: DEBUG nova.virt.hardware [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 601.229099] env[63021]: DEBUG nova.virt.hardware [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 601.229295] env[63021]: DEBUG nova.virt.hardware [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 601.229437] env[63021]: DEBUG nova.virt.hardware [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 601.229580] env[63021]: DEBUG nova.virt.hardware [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 601.230019] env[63021]: DEBUG nova.virt.hardware [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 601.230190] env[63021]: DEBUG nova.virt.hardware [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 601.230349] env[63021]: DEBUG nova.virt.hardware [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 601.230504] env[63021]: DEBUG nova.virt.hardware [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 601.230680] env[63021]: DEBUG nova.virt.hardware [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 601.232121] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcb99b07-8632-405e-81cf-6129f5bd5066 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.236952] env[63021]: DEBUG nova.network.neutron [req-a1ec9292-2bbe-4727-ad80-a140e34de000 req-d2ef1d10-27a1-491f-858c-de93eeaa97cd service nova] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 601.245147] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b596c86-9e04-4c9a-99c8-a850e01f8fd4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.280771] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.118s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 601.281478] env[63021]: ERROR nova.compute.manager [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 40185534-7f2c-4524-98aa-8dad833d2c43, please check neutron logs for more information. [ 601.281478] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Traceback (most recent call last): [ 601.281478] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 601.281478] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] self.driver.spawn(context, instance, image_meta, [ 601.281478] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 601.281478] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 601.281478] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 601.281478] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] vm_ref = self.build_virtual_machine(instance, [ 601.281478] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 601.281478] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] vif_infos = vmwarevif.get_vif_info(self._session, [ 601.281478] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 601.281919] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] for vif in network_info: [ 601.281919] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 601.281919] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] return self._sync_wrapper(fn, *args, **kwargs) [ 601.281919] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 601.281919] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] self.wait() [ 601.281919] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 601.281919] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] self[:] = self._gt.wait() [ 601.281919] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 601.281919] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] return self._exit_event.wait() [ 601.281919] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 601.281919] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] result = hub.switch() [ 601.281919] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 601.281919] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] return self.greenlet.switch() [ 601.282369] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.282369] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] result = function(*args, **kwargs) [ 601.282369] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 601.282369] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] return func(*args, **kwargs) [ 601.282369] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 601.282369] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] raise e [ 601.282369] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.282369] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] nwinfo = self.network_api.allocate_for_instance( [ 601.282369] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.282369] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] created_port_ids = self._update_ports_for_instance( [ 601.282369] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.282369] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] with excutils.save_and_reraise_exception(): [ 601.282369] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.282786] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] self.force_reraise() [ 601.282786] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.282786] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] raise self.value [ 601.282786] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.282786] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] updated_port = self._update_port( [ 601.282786] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.282786] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] _ensure_no_port_binding_failure(port) [ 601.282786] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.282786] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] raise exception.PortBindingFailed(port_id=port['id']) [ 601.282786] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] nova.exception.PortBindingFailed: Binding failed for port 40185534-7f2c-4524-98aa-8dad833d2c43, please check neutron logs for more information. [ 601.282786] env[63021]: ERROR nova.compute.manager [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] [ 601.283148] env[63021]: DEBUG nova.compute.utils [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Binding failed for port 40185534-7f2c-4524-98aa-8dad833d2c43, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 601.283402] env[63021]: DEBUG oslo_concurrency.lockutils [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.828s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 601.284920] env[63021]: INFO nova.compute.claims [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 601.287804] env[63021]: DEBUG nova.compute.manager [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Build of instance ba8e47f7-e10d-439a-9a5a-714655a29dc0 was re-scheduled: Binding failed for port 40185534-7f2c-4524-98aa-8dad833d2c43, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 601.288254] env[63021]: DEBUG nova.compute.manager [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 601.288467] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Acquiring lock "refresh_cache-ba8e47f7-e10d-439a-9a5a-714655a29dc0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.288607] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Acquired lock "refresh_cache-ba8e47f7-e10d-439a-9a5a-714655a29dc0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.288755] env[63021]: DEBUG nova.network.neutron [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 601.295792] env[63021]: INFO nova.compute.manager [-] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Took 1.03 seconds to deallocate network for instance. [ 601.302457] env[63021]: DEBUG nova.compute.claims [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 601.302637] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.388442] env[63021]: DEBUG nova.network.neutron [req-a1ec9292-2bbe-4727-ad80-a140e34de000 req-d2ef1d10-27a1-491f-858c-de93eeaa97cd service nova] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.605666] env[63021]: DEBUG nova.network.neutron [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 601.626965] env[63021]: ERROR nova.compute.manager [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5522f00b-405a-4de6-9f40-35ad30faab5c, please check neutron logs for more information. [ 601.626965] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 601.626965] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.626965] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 601.626965] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.626965] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 601.626965] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.626965] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 601.626965] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.626965] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 601.626965] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.626965] env[63021]: ERROR nova.compute.manager raise self.value [ 601.626965] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.626965] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 601.626965] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.626965] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 601.627745] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.627745] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 601.627745] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5522f00b-405a-4de6-9f40-35ad30faab5c, please check neutron logs for more information. [ 601.627745] env[63021]: ERROR nova.compute.manager [ 601.627745] env[63021]: Traceback (most recent call last): [ 601.627745] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 601.627745] env[63021]: listener.cb(fileno) [ 601.627745] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.627745] env[63021]: result = function(*args, **kwargs) [ 601.627745] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 601.627745] env[63021]: return func(*args, **kwargs) [ 601.627745] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 601.627745] env[63021]: raise e [ 601.627745] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.627745] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 601.627745] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.627745] env[63021]: created_port_ids = self._update_ports_for_instance( [ 601.627745] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.627745] env[63021]: with excutils.save_and_reraise_exception(): [ 601.627745] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.627745] env[63021]: self.force_reraise() [ 601.627745] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.627745] env[63021]: raise self.value [ 601.627745] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.627745] env[63021]: updated_port = self._update_port( [ 601.627745] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.627745] env[63021]: _ensure_no_port_binding_failure(port) [ 601.627745] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.627745] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 601.628730] env[63021]: nova.exception.PortBindingFailed: Binding failed for port 5522f00b-405a-4de6-9f40-35ad30faab5c, please check neutron logs for more information. [ 601.628730] env[63021]: Removing descriptor: 17 [ 601.628730] env[63021]: ERROR nova.compute.manager [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5522f00b-405a-4de6-9f40-35ad30faab5c, please check neutron logs for more information. [ 601.628730] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Traceback (most recent call last): [ 601.628730] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 601.628730] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] yield resources [ 601.628730] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 601.628730] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] self.driver.spawn(context, instance, image_meta, [ 601.628730] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 601.628730] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 601.628730] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 601.628730] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] vm_ref = self.build_virtual_machine(instance, [ 601.629174] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 601.629174] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] vif_infos = vmwarevif.get_vif_info(self._session, [ 601.629174] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 601.629174] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] for vif in network_info: [ 601.629174] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 601.629174] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] return self._sync_wrapper(fn, *args, **kwargs) [ 601.629174] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 601.629174] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] self.wait() [ 601.629174] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 601.629174] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] self[:] = self._gt.wait() [ 601.629174] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 601.629174] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] return self._exit_event.wait() [ 601.629174] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 601.629593] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] result = hub.switch() [ 601.629593] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 601.629593] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] return self.greenlet.switch() [ 601.629593] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.629593] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] result = function(*args, **kwargs) [ 601.629593] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 601.629593] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] return func(*args, **kwargs) [ 601.629593] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 601.629593] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] raise e [ 601.629593] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.629593] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] nwinfo = self.network_api.allocate_for_instance( [ 601.629593] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.629593] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] created_port_ids = self._update_ports_for_instance( [ 601.631166] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.631166] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] with excutils.save_and_reraise_exception(): [ 601.631166] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.631166] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] self.force_reraise() [ 601.631166] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.631166] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] raise self.value [ 601.631166] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.631166] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] updated_port = self._update_port( [ 601.631166] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.631166] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] _ensure_no_port_binding_failure(port) [ 601.631166] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.631166] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] raise exception.PortBindingFailed(port_id=port['id']) [ 601.632621] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] nova.exception.PortBindingFailed: Binding failed for port 5522f00b-405a-4de6-9f40-35ad30faab5c, please check neutron logs for more information. [ 601.632621] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] [ 601.632621] env[63021]: INFO nova.compute.manager [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Terminating instance [ 601.632621] env[63021]: DEBUG oslo_concurrency.lockutils [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Acquiring lock "refresh_cache-ee9bdcd3-03ac-4093-bf21-877a2a8847a0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.632621] env[63021]: DEBUG oslo_concurrency.lockutils [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Acquired lock "refresh_cache-ee9bdcd3-03ac-4093-bf21-877a2a8847a0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.632621] env[63021]: DEBUG nova.network.neutron [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 601.654985] env[63021]: DEBUG nova.network.neutron [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.810391] env[63021]: DEBUG nova.network.neutron [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 601.874745] env[63021]: DEBUG nova.network.neutron [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.894310] env[63021]: DEBUG oslo_concurrency.lockutils [req-a1ec9292-2bbe-4727-ad80-a140e34de000 req-d2ef1d10-27a1-491f-858c-de93eeaa97cd service nova] Releasing lock "refresh_cache-1fa1c7c7-b9fe-4e60-8306-3c263c517c28" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 601.894949] env[63021]: DEBUG nova.compute.manager [req-a1ec9292-2bbe-4727-ad80-a140e34de000 req-d2ef1d10-27a1-491f-858c-de93eeaa97cd service nova] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Received event network-vif-deleted-663a49a6-fb86-4199-b1f0-fbf8f32e806c {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 602.154945] env[63021]: DEBUG nova.network.neutron [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.156886] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Releasing lock "refresh_cache-9d474bd0-e825-49f2-9715-14d85495718e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 602.157270] env[63021]: DEBUG nova.compute.manager [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 602.157455] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 602.158299] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ea82f95-32f5-4ea5-a96c-78b0dae86398 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.166579] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 602.166814] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d03f1942-a236-4b97-afda-372c9124ad7c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.172524] env[63021]: DEBUG oslo_vmware.api [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Waiting for the task: (returnval){ [ 602.172524] env[63021]: value = "task-1293573" [ 602.172524] env[63021]: _type = "Task" [ 602.172524] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 602.180211] env[63021]: DEBUG oslo_vmware.api [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Task: {'id': task-1293573, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.290757] env[63021]: DEBUG nova.network.neutron [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.379576] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Releasing lock "refresh_cache-ba8e47f7-e10d-439a-9a5a-714655a29dc0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 602.379804] env[63021]: DEBUG nova.compute.manager [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 602.379963] env[63021]: DEBUG nova.compute.manager [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 602.380147] env[63021]: DEBUG nova.network.neutron [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 602.398491] env[63021]: DEBUG nova.network.neutron [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.685175] env[63021]: DEBUG oslo_vmware.api [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Task: {'id': task-1293573, 'name': PowerOffVM_Task, 'duration_secs': 0.125614} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 602.685474] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 602.685662] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 602.686278] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2863cf4d-5059-44b9-bd20-2ab602511095 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.706891] env[63021]: DEBUG nova.compute.manager [req-684843b7-4013-4661-8a0f-946ac03f0107 req-b5949157-6597-4047-9ca6-59378d4a4c82 service nova] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Received event network-changed-5522f00b-405a-4de6-9f40-35ad30faab5c {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 602.706891] env[63021]: DEBUG nova.compute.manager [req-684843b7-4013-4661-8a0f-946ac03f0107 req-b5949157-6597-4047-9ca6-59378d4a4c82 service nova] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Refreshing instance network info cache due to event network-changed-5522f00b-405a-4de6-9f40-35ad30faab5c. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 602.707058] env[63021]: DEBUG oslo_concurrency.lockutils [req-684843b7-4013-4661-8a0f-946ac03f0107 req-b5949157-6597-4047-9ca6-59378d4a4c82 service nova] Acquiring lock "refresh_cache-ee9bdcd3-03ac-4093-bf21-877a2a8847a0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.711837] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 602.712202] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Deleting contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 602.712416] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Deleting the datastore file [datastore2] 9d474bd0-e825-49f2-9715-14d85495718e {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 602.714963] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bc2b2c41-52f1-4d28-97e0-1b6cb4afb90d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.721865] env[63021]: DEBUG oslo_vmware.api [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Waiting for the task: (returnval){ [ 602.721865] env[63021]: value = "task-1293575" [ 602.721865] env[63021]: _type = "Task" [ 602.721865] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 602.732173] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeaec342-6bb1-4cbe-82de-1348cb899f47 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.734761] env[63021]: DEBUG oslo_vmware.api [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Task: {'id': task-1293575, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.738995] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d0910c6-cc2f-4f55-8d65-5436a76ea316 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.771594] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f6c66ee-8083-473e-8727-30d6511f04a0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.780569] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-583a8c20-835a-48ee-b29c-47847009eb1b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.793563] env[63021]: DEBUG nova.compute.provider_tree [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 602.794858] env[63021]: DEBUG oslo_concurrency.lockutils [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Releasing lock "refresh_cache-ee9bdcd3-03ac-4093-bf21-877a2a8847a0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 602.795229] env[63021]: DEBUG nova.compute.manager [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 602.795631] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 602.796128] env[63021]: DEBUG oslo_concurrency.lockutils [req-684843b7-4013-4661-8a0f-946ac03f0107 req-b5949157-6597-4047-9ca6-59378d4a4c82 service nova] Acquired lock "refresh_cache-ee9bdcd3-03ac-4093-bf21-877a2a8847a0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.796318] env[63021]: DEBUG nova.network.neutron [req-684843b7-4013-4661-8a0f-946ac03f0107 req-b5949157-6597-4047-9ca6-59378d4a4c82 service nova] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Refreshing network info cache for port 5522f00b-405a-4de6-9f40-35ad30faab5c {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 602.797198] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-677070e6-3e77-4d3d-a842-4ccb52bb1d3a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.807174] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e34b1fc-e9ea-4628-9690-7a652ad2620d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.830312] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ee9bdcd3-03ac-4093-bf21-877a2a8847a0 could not be found. [ 602.830586] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 602.830840] env[63021]: INFO nova.compute.manager [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Took 0.04 seconds to destroy the instance on the hypervisor. [ 602.831100] env[63021]: DEBUG oslo.service.loopingcall [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 602.831417] env[63021]: DEBUG nova.compute.manager [-] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 602.831693] env[63021]: DEBUG nova.network.neutron [-] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 602.846989] env[63021]: DEBUG nova.network.neutron [-] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.903693] env[63021]: DEBUG nova.network.neutron [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.232349] env[63021]: DEBUG oslo_vmware.api [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Task: {'id': task-1293575, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.095771} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.233037] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 603.233037] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Deleted contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 603.233204] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 603.233391] env[63021]: INFO nova.compute.manager [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Took 1.08 seconds to destroy the instance on the hypervisor. [ 603.233632] env[63021]: DEBUG oslo.service.loopingcall [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 603.233813] env[63021]: DEBUG nova.compute.manager [-] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 603.233905] env[63021]: DEBUG nova.network.neutron [-] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 603.249445] env[63021]: DEBUG nova.network.neutron [-] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.297263] env[63021]: DEBUG nova.scheduler.client.report [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 603.321943] env[63021]: DEBUG nova.network.neutron [req-684843b7-4013-4661-8a0f-946ac03f0107 req-b5949157-6597-4047-9ca6-59378d4a4c82 service nova] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.351832] env[63021]: DEBUG nova.network.neutron [-] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.398731] env[63021]: DEBUG nova.network.neutron [req-684843b7-4013-4661-8a0f-946ac03f0107 req-b5949157-6597-4047-9ca6-59378d4a4c82 service nova] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.406366] env[63021]: INFO nova.compute.manager [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] [instance: ba8e47f7-e10d-439a-9a5a-714655a29dc0] Took 1.03 seconds to deallocate network for instance. [ 603.751698] env[63021]: DEBUG nova.network.neutron [-] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.804973] env[63021]: DEBUG oslo_concurrency.lockutils [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.521s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.805486] env[63021]: DEBUG nova.compute.manager [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 603.810488] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 15.542s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.810738] env[63021]: DEBUG nova.objects.instance [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63021) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 603.854618] env[63021]: INFO nova.compute.manager [-] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Took 1.02 seconds to deallocate network for instance. [ 603.857117] env[63021]: DEBUG nova.compute.claims [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 603.857309] env[63021]: DEBUG oslo_concurrency.lockutils [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.901925] env[63021]: DEBUG oslo_concurrency.lockutils [req-684843b7-4013-4661-8a0f-946ac03f0107 req-b5949157-6597-4047-9ca6-59378d4a4c82 service nova] Releasing lock "refresh_cache-ee9bdcd3-03ac-4093-bf21-877a2a8847a0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.902211] env[63021]: DEBUG nova.compute.manager [req-684843b7-4013-4661-8a0f-946ac03f0107 req-b5949157-6597-4047-9ca6-59378d4a4c82 service nova] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Received event network-vif-deleted-5522f00b-405a-4de6-9f40-35ad30faab5c {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 604.256150] env[63021]: INFO nova.compute.manager [-] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Took 1.02 seconds to deallocate network for instance. [ 604.313047] env[63021]: DEBUG nova.compute.utils [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 604.313586] env[63021]: DEBUG nova.compute.manager [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 604.314193] env[63021]: DEBUG nova.network.neutron [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 604.354513] env[63021]: DEBUG nova.policy [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '25cca50a45ae49ccba6409fa98b21844', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7915bdde147547009ea38eea425de28e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 604.438924] env[63021]: INFO nova.scheduler.client.report [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Deleted allocations for instance ba8e47f7-e10d-439a-9a5a-714655a29dc0 [ 604.625672] env[63021]: DEBUG nova.network.neutron [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Successfully created port: 36395c0e-99a7-4f1c-837f-5ed5718b3abc {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 604.761951] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.819132] env[63021]: DEBUG nova.compute.manager [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 604.822786] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f7b3320e-fe82-4c57-90b9-7a527d1b69eb tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.826317] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.318s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.826317] env[63021]: INFO nova.compute.claims [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 604.949027] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c47c1df1-84c2-4590-89c3-ccb805c8d1de tempest-ListImageFiltersTestJSON-889288412 tempest-ListImageFiltersTestJSON-889288412-project-member] Lock "ba8e47f7-e10d-439a-9a5a-714655a29dc0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.303s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.424307] env[63021]: DEBUG nova.compute.manager [req-cbe4ca81-0731-4274-ac28-5a1cb2c0a0f7 req-5587c567-f114-416a-a7ee-741a6c418db2 service nova] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Received event network-changed-36395c0e-99a7-4f1c-837f-5ed5718b3abc {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 605.424900] env[63021]: DEBUG nova.compute.manager [req-cbe4ca81-0731-4274-ac28-5a1cb2c0a0f7 req-5587c567-f114-416a-a7ee-741a6c418db2 service nova] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Refreshing instance network info cache due to event network-changed-36395c0e-99a7-4f1c-837f-5ed5718b3abc. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 605.425296] env[63021]: DEBUG oslo_concurrency.lockutils [req-cbe4ca81-0731-4274-ac28-5a1cb2c0a0f7 req-5587c567-f114-416a-a7ee-741a6c418db2 service nova] Acquiring lock "refresh_cache-e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.425718] env[63021]: DEBUG oslo_concurrency.lockutils [req-cbe4ca81-0731-4274-ac28-5a1cb2c0a0f7 req-5587c567-f114-416a-a7ee-741a6c418db2 service nova] Acquired lock "refresh_cache-e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.425971] env[63021]: DEBUG nova.network.neutron [req-cbe4ca81-0731-4274-ac28-5a1cb2c0a0f7 req-5587c567-f114-416a-a7ee-741a6c418db2 service nova] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Refreshing network info cache for port 36395c0e-99a7-4f1c-837f-5ed5718b3abc {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 605.456020] env[63021]: DEBUG nova.compute.manager [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 605.517851] env[63021]: ERROR nova.compute.manager [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 36395c0e-99a7-4f1c-837f-5ed5718b3abc, please check neutron logs for more information. [ 605.517851] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 605.517851] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 605.517851] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 605.517851] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 605.517851] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 605.517851] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 605.517851] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 605.517851] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.517851] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 605.517851] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.517851] env[63021]: ERROR nova.compute.manager raise self.value [ 605.517851] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 605.517851] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 605.517851] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.517851] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 605.518385] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.518385] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 605.518385] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 36395c0e-99a7-4f1c-837f-5ed5718b3abc, please check neutron logs for more information. [ 605.518385] env[63021]: ERROR nova.compute.manager [ 605.518385] env[63021]: Traceback (most recent call last): [ 605.518385] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 605.518385] env[63021]: listener.cb(fileno) [ 605.518385] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.518385] env[63021]: result = function(*args, **kwargs) [ 605.518385] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 605.518385] env[63021]: return func(*args, **kwargs) [ 605.518385] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 605.518385] env[63021]: raise e [ 605.518385] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 605.518385] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 605.518385] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 605.518385] env[63021]: created_port_ids = self._update_ports_for_instance( [ 605.518385] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 605.518385] env[63021]: with excutils.save_and_reraise_exception(): [ 605.518385] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.518385] env[63021]: self.force_reraise() [ 605.518385] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.518385] env[63021]: raise self.value [ 605.518385] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 605.518385] env[63021]: updated_port = self._update_port( [ 605.518385] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.518385] env[63021]: _ensure_no_port_binding_failure(port) [ 605.518385] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.518385] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 605.519284] env[63021]: nova.exception.PortBindingFailed: Binding failed for port 36395c0e-99a7-4f1c-837f-5ed5718b3abc, please check neutron logs for more information. [ 605.519284] env[63021]: Removing descriptor: 15 [ 605.834800] env[63021]: DEBUG nova.compute.manager [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 605.867688] env[63021]: DEBUG nova.virt.hardware [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 605.867688] env[63021]: DEBUG nova.virt.hardware [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 605.867688] env[63021]: DEBUG nova.virt.hardware [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 605.867854] env[63021]: DEBUG nova.virt.hardware [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 605.867854] env[63021]: DEBUG nova.virt.hardware [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 605.867854] env[63021]: DEBUG nova.virt.hardware [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 605.867854] env[63021]: DEBUG nova.virt.hardware [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 605.868255] env[63021]: DEBUG nova.virt.hardware [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 605.868665] env[63021]: DEBUG nova.virt.hardware [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 605.868960] env[63021]: DEBUG nova.virt.hardware [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 605.869264] env[63021]: DEBUG nova.virt.hardware [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 605.870452] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0e4bd41-5064-4bab-a952-b7e9b08d9984 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.885115] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c1f44c1-8fcf-49d4-a653-3144df752487 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.899992] env[63021]: ERROR nova.compute.manager [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 36395c0e-99a7-4f1c-837f-5ed5718b3abc, please check neutron logs for more information. [ 605.899992] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Traceback (most recent call last): [ 605.899992] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 605.899992] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] yield resources [ 605.899992] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 605.899992] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] self.driver.spawn(context, instance, image_meta, [ 605.899992] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 605.899992] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 605.899992] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 605.899992] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] vm_ref = self.build_virtual_machine(instance, [ 605.899992] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 605.900433] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] vif_infos = vmwarevif.get_vif_info(self._session, [ 605.900433] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 605.900433] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] for vif in network_info: [ 605.900433] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 605.900433] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] return self._sync_wrapper(fn, *args, **kwargs) [ 605.900433] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 605.900433] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] self.wait() [ 605.900433] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 605.900433] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] self[:] = self._gt.wait() [ 605.900433] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 605.900433] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] return self._exit_event.wait() [ 605.900433] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 605.900433] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] current.throw(*self._exc) [ 605.900848] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.900848] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] result = function(*args, **kwargs) [ 605.900848] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 605.900848] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] return func(*args, **kwargs) [ 605.900848] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 605.900848] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] raise e [ 605.900848] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 605.900848] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] nwinfo = self.network_api.allocate_for_instance( [ 605.900848] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 605.900848] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] created_port_ids = self._update_ports_for_instance( [ 605.900848] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 605.900848] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] with excutils.save_and_reraise_exception(): [ 605.900848] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.901383] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] self.force_reraise() [ 605.901383] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.901383] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] raise self.value [ 605.901383] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 605.901383] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] updated_port = self._update_port( [ 605.901383] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.901383] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] _ensure_no_port_binding_failure(port) [ 605.901383] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.901383] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] raise exception.PortBindingFailed(port_id=port['id']) [ 605.901383] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] nova.exception.PortBindingFailed: Binding failed for port 36395c0e-99a7-4f1c-837f-5ed5718b3abc, please check neutron logs for more information. [ 605.901383] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] [ 605.901383] env[63021]: INFO nova.compute.manager [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Terminating instance [ 605.904391] env[63021]: DEBUG oslo_concurrency.lockutils [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Acquiring lock "refresh_cache-e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.953129] env[63021]: DEBUG nova.network.neutron [req-cbe4ca81-0731-4274-ac28-5a1cb2c0a0f7 req-5587c567-f114-416a-a7ee-741a6c418db2 service nova] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 605.979826] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.049057] env[63021]: DEBUG nova.network.neutron [req-cbe4ca81-0731-4274-ac28-5a1cb2c0a0f7 req-5587c567-f114-416a-a7ee-741a6c418db2 service nova] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.268604] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5c682c1-52b7-425e-a7d5-9d92fbff3e16 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.277769] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74513618-002b-4a9e-a549-07f1e8e93b47 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.308037] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6783dc75-40bb-439a-83c4-b68caf27c08b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.315288] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b61d962-68e7-4fc3-bb03-d12ee2dc8195 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.328905] env[63021]: DEBUG nova.compute.provider_tree [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Updating inventory in ProviderTree for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 606.551901] env[63021]: DEBUG oslo_concurrency.lockutils [req-cbe4ca81-0731-4274-ac28-5a1cb2c0a0f7 req-5587c567-f114-416a-a7ee-741a6c418db2 service nova] Releasing lock "refresh_cache-e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.552360] env[63021]: DEBUG oslo_concurrency.lockutils [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Acquired lock "refresh_cache-e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.552541] env[63021]: DEBUG nova.network.neutron [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 606.850913] env[63021]: ERROR nova.scheduler.client.report [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [req-0b418d2e-5bea-405d-91d3-72731912714f] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 6ef6c041-2b0d-4bc3-8662-5d679a77a86d. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-0b418d2e-5bea-405d-91d3-72731912714f"}]} [ 606.869217] env[63021]: DEBUG nova.scheduler.client.report [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Refreshing inventories for resource provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 606.883394] env[63021]: DEBUG nova.scheduler.client.report [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Updating ProviderTree inventory for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 606.883642] env[63021]: DEBUG nova.compute.provider_tree [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Updating inventory in ProviderTree for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 606.895936] env[63021]: DEBUG nova.scheduler.client.report [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Refreshing aggregate associations for resource provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d, aggregates: None {{(pid=63021) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 606.914230] env[63021]: DEBUG nova.scheduler.client.report [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Refreshing trait associations for resource provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE {{(pid=63021) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 607.073797] env[63021]: DEBUG nova.network.neutron [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 607.193343] env[63021]: DEBUG nova.network.neutron [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.287590] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f87a706-5983-48c4-ab44-e6c65f1e5177 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.295028] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b564beed-9e09-4eec-a84b-ce6c61a9bfde {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.324403] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-694fea1d-65ac-4fa6-9e74-f3d78cd1ddcf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.331623] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3e91d1b-14c5-4d17-ac63-0c366b374b19 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.346304] env[63021]: DEBUG nova.compute.provider_tree [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Updating inventory in ProviderTree for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 607.506492] env[63021]: DEBUG nova.compute.manager [req-890c933b-0737-4d3d-aaea-5455fec50b58 req-f6e6a99c-14ac-4733-be4b-df0abb705d50 service nova] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Received event network-vif-deleted-36395c0e-99a7-4f1c-837f-5ed5718b3abc {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 607.696597] env[63021]: DEBUG oslo_concurrency.lockutils [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Releasing lock "refresh_cache-e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 607.697144] env[63021]: DEBUG nova.compute.manager [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 607.697330] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 607.697722] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-31c1c0ed-b89c-42aa-95cb-ad9678ad21a7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.708367] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07a56bab-25b1-42c2-8e67-147fdc6ca97d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.730948] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7 could not be found. [ 607.731217] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 607.731402] env[63021]: INFO nova.compute.manager [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Took 0.03 seconds to destroy the instance on the hypervisor. [ 607.731632] env[63021]: DEBUG oslo.service.loopingcall [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 607.731840] env[63021]: DEBUG nova.compute.manager [-] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 607.731935] env[63021]: DEBUG nova.network.neutron [-] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 607.752199] env[63021]: DEBUG nova.network.neutron [-] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 607.889617] env[63021]: DEBUG nova.scheduler.client.report [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Updated inventory for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with generation 46 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 607.889979] env[63021]: DEBUG nova.compute.provider_tree [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Updating resource provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d generation from 46 to 47 during operation: update_inventory {{(pid=63021) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 607.890229] env[63021]: DEBUG nova.compute.provider_tree [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Updating inventory in ProviderTree for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 608.255187] env[63021]: DEBUG nova.network.neutron [-] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.398999] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.574s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.399517] env[63021]: DEBUG nova.compute.manager [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 608.403081] env[63021]: DEBUG oslo_concurrency.lockutils [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.714s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.404825] env[63021]: INFO nova.compute.claims [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 608.757082] env[63021]: INFO nova.compute.manager [-] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Took 1.02 seconds to deallocate network for instance. [ 608.760281] env[63021]: DEBUG nova.compute.claims [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 608.760281] env[63021]: DEBUG oslo_concurrency.lockutils [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.909130] env[63021]: DEBUG nova.compute.utils [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 608.912986] env[63021]: DEBUG nova.compute.manager [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 608.912986] env[63021]: DEBUG nova.network.neutron [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 608.966576] env[63021]: DEBUG nova.policy [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2642402689b04f89946802ac3b3e6287', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '805af233efa147b3982fb794cd9e6f73', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 609.267835] env[63021]: DEBUG nova.network.neutron [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Successfully created port: b5fe80a8-6652-498a-898b-2ed43886c4fd {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 609.414233] env[63021]: DEBUG nova.compute.manager [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 609.859609] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16491c50-2a97-4340-96a7-e70b2569ff64 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.867973] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4586b8e-a80c-42c4-b267-42af3b1f69d7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.900960] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc4777c5-5873-4b1e-ab62-ab9dcc232ca9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.908649] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19b257c7-f0ce-40a0-8292-f78c7226deaf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.926420] env[63021]: DEBUG nova.compute.provider_tree [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 610.118781] env[63021]: DEBUG nova.compute.manager [req-d561e4b0-0294-4e6a-bbdb-8d0646ce6684 req-83723865-372c-42ff-a904-f82b1c0ce090 service nova] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Received event network-changed-b5fe80a8-6652-498a-898b-2ed43886c4fd {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 610.119017] env[63021]: DEBUG nova.compute.manager [req-d561e4b0-0294-4e6a-bbdb-8d0646ce6684 req-83723865-372c-42ff-a904-f82b1c0ce090 service nova] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Refreshing instance network info cache due to event network-changed-b5fe80a8-6652-498a-898b-2ed43886c4fd. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 610.119229] env[63021]: DEBUG oslo_concurrency.lockutils [req-d561e4b0-0294-4e6a-bbdb-8d0646ce6684 req-83723865-372c-42ff-a904-f82b1c0ce090 service nova] Acquiring lock "refresh_cache-17ff6693-c323-478b-8b61-776c3b41f77e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.119356] env[63021]: DEBUG oslo_concurrency.lockutils [req-d561e4b0-0294-4e6a-bbdb-8d0646ce6684 req-83723865-372c-42ff-a904-f82b1c0ce090 service nova] Acquired lock "refresh_cache-17ff6693-c323-478b-8b61-776c3b41f77e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.119515] env[63021]: DEBUG nova.network.neutron [req-d561e4b0-0294-4e6a-bbdb-8d0646ce6684 req-83723865-372c-42ff-a904-f82b1c0ce090 service nova] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Refreshing network info cache for port b5fe80a8-6652-498a-898b-2ed43886c4fd {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 610.174219] env[63021]: ERROR nova.compute.manager [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b5fe80a8-6652-498a-898b-2ed43886c4fd, please check neutron logs for more information. [ 610.174219] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 610.174219] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.174219] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 610.174219] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.174219] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 610.174219] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.174219] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 610.174219] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.174219] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 610.174219] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.174219] env[63021]: ERROR nova.compute.manager raise self.value [ 610.174219] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.174219] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 610.174219] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.174219] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 610.174730] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.174730] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 610.174730] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b5fe80a8-6652-498a-898b-2ed43886c4fd, please check neutron logs for more information. [ 610.174730] env[63021]: ERROR nova.compute.manager [ 610.174730] env[63021]: Traceback (most recent call last): [ 610.174730] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 610.174730] env[63021]: listener.cb(fileno) [ 610.174730] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.174730] env[63021]: result = function(*args, **kwargs) [ 610.174730] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 610.174730] env[63021]: return func(*args, **kwargs) [ 610.174730] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.174730] env[63021]: raise e [ 610.174730] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.174730] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 610.174730] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.174730] env[63021]: created_port_ids = self._update_ports_for_instance( [ 610.174730] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.174730] env[63021]: with excutils.save_and_reraise_exception(): [ 610.174730] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.174730] env[63021]: self.force_reraise() [ 610.174730] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.174730] env[63021]: raise self.value [ 610.174730] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.174730] env[63021]: updated_port = self._update_port( [ 610.174730] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.174730] env[63021]: _ensure_no_port_binding_failure(port) [ 610.174730] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.174730] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 610.175718] env[63021]: nova.exception.PortBindingFailed: Binding failed for port b5fe80a8-6652-498a-898b-2ed43886c4fd, please check neutron logs for more information. [ 610.175718] env[63021]: Removing descriptor: 15 [ 610.432724] env[63021]: DEBUG nova.compute.manager [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 610.435681] env[63021]: DEBUG nova.scheduler.client.report [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 610.469401] env[63021]: DEBUG nova.virt.hardware [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 610.469648] env[63021]: DEBUG nova.virt.hardware [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 610.469799] env[63021]: DEBUG nova.virt.hardware [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 610.469972] env[63021]: DEBUG nova.virt.hardware [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 610.470133] env[63021]: DEBUG nova.virt.hardware [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 610.470276] env[63021]: DEBUG nova.virt.hardware [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 610.470477] env[63021]: DEBUG nova.virt.hardware [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 610.470628] env[63021]: DEBUG nova.virt.hardware [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 610.470787] env[63021]: DEBUG nova.virt.hardware [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 610.470944] env[63021]: DEBUG nova.virt.hardware [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 610.471338] env[63021]: DEBUG nova.virt.hardware [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 610.472445] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4969e9b1-02dd-41d7-bea4-c3ea2b77643d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.480451] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2850d4cd-c08d-44a6-99ad-9d311fb50023 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.500486] env[63021]: ERROR nova.compute.manager [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b5fe80a8-6652-498a-898b-2ed43886c4fd, please check neutron logs for more information. [ 610.500486] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Traceback (most recent call last): [ 610.500486] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 610.500486] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] yield resources [ 610.500486] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 610.500486] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] self.driver.spawn(context, instance, image_meta, [ 610.500486] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 610.500486] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 610.500486] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 610.500486] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] vm_ref = self.build_virtual_machine(instance, [ 610.500486] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 610.500888] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] vif_infos = vmwarevif.get_vif_info(self._session, [ 610.500888] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 610.500888] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] for vif in network_info: [ 610.500888] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 610.500888] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] return self._sync_wrapper(fn, *args, **kwargs) [ 610.500888] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 610.500888] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] self.wait() [ 610.500888] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 610.500888] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] self[:] = self._gt.wait() [ 610.500888] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 610.500888] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] return self._exit_event.wait() [ 610.500888] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 610.500888] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] current.throw(*self._exc) [ 610.501390] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.501390] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] result = function(*args, **kwargs) [ 610.501390] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 610.501390] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] return func(*args, **kwargs) [ 610.501390] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.501390] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] raise e [ 610.501390] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.501390] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] nwinfo = self.network_api.allocate_for_instance( [ 610.501390] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.501390] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] created_port_ids = self._update_ports_for_instance( [ 610.501390] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.501390] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] with excutils.save_and_reraise_exception(): [ 610.501390] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.501758] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] self.force_reraise() [ 610.501758] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.501758] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] raise self.value [ 610.501758] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.501758] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] updated_port = self._update_port( [ 610.501758] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.501758] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] _ensure_no_port_binding_failure(port) [ 610.501758] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.501758] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] raise exception.PortBindingFailed(port_id=port['id']) [ 610.501758] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] nova.exception.PortBindingFailed: Binding failed for port b5fe80a8-6652-498a-898b-2ed43886c4fd, please check neutron logs for more information. [ 610.501758] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] [ 610.501758] env[63021]: INFO nova.compute.manager [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Terminating instance [ 610.502853] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "refresh_cache-17ff6693-c323-478b-8b61-776c3b41f77e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.641018] env[63021]: DEBUG nova.network.neutron [req-d561e4b0-0294-4e6a-bbdb-8d0646ce6684 req-83723865-372c-42ff-a904-f82b1c0ce090 service nova] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 610.722030] env[63021]: DEBUG nova.network.neutron [req-d561e4b0-0294-4e6a-bbdb-8d0646ce6684 req-83723865-372c-42ff-a904-f82b1c0ce090 service nova] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.942358] env[63021]: DEBUG oslo_concurrency.lockutils [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.540s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.942900] env[63021]: DEBUG nova.compute.manager [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 610.946118] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.269s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.947708] env[63021]: INFO nova.compute.claims [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 611.224999] env[63021]: DEBUG oslo_concurrency.lockutils [req-d561e4b0-0294-4e6a-bbdb-8d0646ce6684 req-83723865-372c-42ff-a904-f82b1c0ce090 service nova] Releasing lock "refresh_cache-17ff6693-c323-478b-8b61-776c3b41f77e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.225900] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquired lock "refresh_cache-17ff6693-c323-478b-8b61-776c3b41f77e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.225900] env[63021]: DEBUG nova.network.neutron [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 611.452963] env[63021]: DEBUG nova.compute.utils [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 611.456272] env[63021]: DEBUG nova.compute.manager [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 611.456446] env[63021]: DEBUG nova.network.neutron [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 611.521166] env[63021]: DEBUG nova.policy [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c468d6ce6b74478eb5adef00f6f05bb7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4df48a4180a943b5911f9d287b86f016', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 611.744288] env[63021]: DEBUG nova.network.neutron [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.783716] env[63021]: DEBUG nova.network.neutron [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Successfully created port: 8b1b2b2c-fec7-4c23-8349-bc89d9cc5b36 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 611.893184] env[63021]: DEBUG nova.network.neutron [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.957379] env[63021]: DEBUG nova.compute.manager [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 612.148271] env[63021]: DEBUG nova.compute.manager [req-55c80131-c8e2-4669-adb2-31fb82b016fa req-8a672e03-e348-4f9b-9185-845eb3f01340 service nova] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Received event network-vif-deleted-b5fe80a8-6652-498a-898b-2ed43886c4fd {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 612.396531] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Releasing lock "refresh_cache-17ff6693-c323-478b-8b61-776c3b41f77e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.396949] env[63021]: DEBUG nova.compute.manager [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 612.397157] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 612.397453] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3682a89f-0469-4db6-b815-2ace0d9bcc87 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.409112] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2d228bc-9b11-4d79-9e44-eecc6ce4b768 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.436498] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 17ff6693-c323-478b-8b61-776c3b41f77e could not be found. [ 612.436735] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 612.436917] env[63021]: INFO nova.compute.manager [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 612.437180] env[63021]: DEBUG oslo.service.loopingcall [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 612.438332] env[63021]: DEBUG nova.compute.manager [-] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 612.438438] env[63021]: DEBUG nova.network.neutron [-] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 612.440582] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-306a65d7-88ad-4dda-bf72-d8a32735a224 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.447353] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18fa2a17-eaba-45ae-8521-957e53ba03ab {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.482939] env[63021]: DEBUG nova.network.neutron [-] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.487137] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de2e7e74-f924-47e8-8108-2bb4203d8290 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.497884] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8030f8e-f34f-44ee-9ec0-8699fbe1d5e0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.514809] env[63021]: DEBUG nova.compute.provider_tree [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 612.711401] env[63021]: ERROR nova.compute.manager [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8b1b2b2c-fec7-4c23-8349-bc89d9cc5b36, please check neutron logs for more information. [ 612.711401] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 612.711401] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 612.711401] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 612.711401] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 612.711401] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 612.711401] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 612.711401] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 612.711401] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 612.711401] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 612.711401] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 612.711401] env[63021]: ERROR nova.compute.manager raise self.value [ 612.711401] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 612.711401] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 612.711401] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 612.711401] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 612.712071] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 612.712071] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 612.712071] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8b1b2b2c-fec7-4c23-8349-bc89d9cc5b36, please check neutron logs for more information. [ 612.712071] env[63021]: ERROR nova.compute.manager [ 612.712071] env[63021]: Traceback (most recent call last): [ 612.712071] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 612.712071] env[63021]: listener.cb(fileno) [ 612.712071] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 612.712071] env[63021]: result = function(*args, **kwargs) [ 612.712071] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 612.712071] env[63021]: return func(*args, **kwargs) [ 612.712071] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 612.712071] env[63021]: raise e [ 612.712071] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 612.712071] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 612.712071] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 612.712071] env[63021]: created_port_ids = self._update_ports_for_instance( [ 612.712071] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 612.712071] env[63021]: with excutils.save_and_reraise_exception(): [ 612.712071] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 612.712071] env[63021]: self.force_reraise() [ 612.712071] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 612.712071] env[63021]: raise self.value [ 612.712071] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 612.712071] env[63021]: updated_port = self._update_port( [ 612.712071] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 612.712071] env[63021]: _ensure_no_port_binding_failure(port) [ 612.712071] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 612.712071] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 612.714667] env[63021]: nova.exception.PortBindingFailed: Binding failed for port 8b1b2b2c-fec7-4c23-8349-bc89d9cc5b36, please check neutron logs for more information. [ 612.714667] env[63021]: Removing descriptor: 15 [ 612.984024] env[63021]: DEBUG nova.compute.manager [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 612.986633] env[63021]: DEBUG nova.network.neutron [-] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.012157] env[63021]: DEBUG nova.virt.hardware [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 613.012590] env[63021]: DEBUG nova.virt.hardware [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 613.012590] env[63021]: DEBUG nova.virt.hardware [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 613.012720] env[63021]: DEBUG nova.virt.hardware [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 613.012845] env[63021]: DEBUG nova.virt.hardware [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 613.012986] env[63021]: DEBUG nova.virt.hardware [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 613.013260] env[63021]: DEBUG nova.virt.hardware [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 613.013437] env[63021]: DEBUG nova.virt.hardware [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 613.013611] env[63021]: DEBUG nova.virt.hardware [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 613.013758] env[63021]: DEBUG nova.virt.hardware [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 613.013922] env[63021]: DEBUG nova.virt.hardware [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 613.014749] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbe4d706-5f4f-4fe7-8d28-3ad115cd00cc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.018417] env[63021]: DEBUG nova.scheduler.client.report [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 613.024426] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fadf4b33-9a7c-4957-8d32-d748c64ba6ae {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.038665] env[63021]: ERROR nova.compute.manager [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8b1b2b2c-fec7-4c23-8349-bc89d9cc5b36, please check neutron logs for more information. [ 613.038665] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Traceback (most recent call last): [ 613.038665] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 613.038665] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] yield resources [ 613.038665] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 613.038665] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] self.driver.spawn(context, instance, image_meta, [ 613.038665] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 613.038665] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 613.038665] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 613.038665] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] vm_ref = self.build_virtual_machine(instance, [ 613.038665] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 613.039157] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] vif_infos = vmwarevif.get_vif_info(self._session, [ 613.039157] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 613.039157] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] for vif in network_info: [ 613.039157] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 613.039157] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] return self._sync_wrapper(fn, *args, **kwargs) [ 613.039157] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 613.039157] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] self.wait() [ 613.039157] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 613.039157] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] self[:] = self._gt.wait() [ 613.039157] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 613.039157] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] return self._exit_event.wait() [ 613.039157] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 613.039157] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] current.throw(*self._exc) [ 613.039623] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 613.039623] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] result = function(*args, **kwargs) [ 613.039623] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 613.039623] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] return func(*args, **kwargs) [ 613.039623] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 613.039623] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] raise e [ 613.039623] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.039623] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] nwinfo = self.network_api.allocate_for_instance( [ 613.039623] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 613.039623] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] created_port_ids = self._update_ports_for_instance( [ 613.039623] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 613.039623] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] with excutils.save_and_reraise_exception(): [ 613.039623] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.040103] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] self.force_reraise() [ 613.040103] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.040103] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] raise self.value [ 613.040103] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 613.040103] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] updated_port = self._update_port( [ 613.040103] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.040103] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] _ensure_no_port_binding_failure(port) [ 613.040103] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.040103] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] raise exception.PortBindingFailed(port_id=port['id']) [ 613.040103] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] nova.exception.PortBindingFailed: Binding failed for port 8b1b2b2c-fec7-4c23-8349-bc89d9cc5b36, please check neutron logs for more information. [ 613.040103] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] [ 613.040103] env[63021]: INFO nova.compute.manager [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Terminating instance [ 613.040759] env[63021]: DEBUG oslo_concurrency.lockutils [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Acquiring lock "refresh_cache-efd0e375-3134-4984-8f82-48b53c65e26c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.040911] env[63021]: DEBUG oslo_concurrency.lockutils [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Acquired lock "refresh_cache-efd0e375-3134-4984-8f82-48b53c65e26c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.041104] env[63021]: DEBUG nova.network.neutron [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 613.489474] env[63021]: INFO nova.compute.manager [-] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Took 1.05 seconds to deallocate network for instance. [ 613.493489] env[63021]: DEBUG nova.compute.claims [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 613.493663] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.524761] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.579s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.525268] env[63021]: DEBUG nova.compute.manager [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 613.527750] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.666s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.529226] env[63021]: INFO nova.compute.claims [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 613.563666] env[63021]: DEBUG nova.network.neutron [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.618536] env[63021]: DEBUG nova.network.neutron [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.033950] env[63021]: DEBUG nova.compute.utils [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 614.037099] env[63021]: DEBUG nova.compute.manager [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 614.037285] env[63021]: DEBUG nova.network.neutron [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 614.082149] env[63021]: DEBUG nova.policy [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '34d716fca4204dbf98ed61c9d8d9e90d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '42fc3b8ae5404d87a2efbd7cdd7be0c7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 614.122650] env[63021]: DEBUG oslo_concurrency.lockutils [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Releasing lock "refresh_cache-efd0e375-3134-4984-8f82-48b53c65e26c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.122902] env[63021]: DEBUG nova.compute.manager [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 614.123110] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 614.123449] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6be022b0-47ab-42ad-9bc0-202d7caa28c1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.133039] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4399cf55-b45f-48d9-85fb-5583a6ca3afc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.157344] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance efd0e375-3134-4984-8f82-48b53c65e26c could not be found. [ 614.157692] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 614.157771] env[63021]: INFO nova.compute.manager [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Took 0.03 seconds to destroy the instance on the hypervisor. [ 614.158184] env[63021]: DEBUG oslo.service.loopingcall [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 614.158268] env[63021]: DEBUG nova.compute.manager [-] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 614.158305] env[63021]: DEBUG nova.network.neutron [-] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 614.173021] env[63021]: DEBUG nova.compute.manager [req-fa8c7c68-2a56-4cb3-b572-64272440557b req-2157dd1b-54d9-4d7e-bc03-e279b7c9d867 service nova] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Received event network-changed-8b1b2b2c-fec7-4c23-8349-bc89d9cc5b36 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 614.173021] env[63021]: DEBUG nova.compute.manager [req-fa8c7c68-2a56-4cb3-b572-64272440557b req-2157dd1b-54d9-4d7e-bc03-e279b7c9d867 service nova] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Refreshing instance network info cache due to event network-changed-8b1b2b2c-fec7-4c23-8349-bc89d9cc5b36. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 614.173183] env[63021]: DEBUG oslo_concurrency.lockutils [req-fa8c7c68-2a56-4cb3-b572-64272440557b req-2157dd1b-54d9-4d7e-bc03-e279b7c9d867 service nova] Acquiring lock "refresh_cache-efd0e375-3134-4984-8f82-48b53c65e26c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.173310] env[63021]: DEBUG oslo_concurrency.lockutils [req-fa8c7c68-2a56-4cb3-b572-64272440557b req-2157dd1b-54d9-4d7e-bc03-e279b7c9d867 service nova] Acquired lock "refresh_cache-efd0e375-3134-4984-8f82-48b53c65e26c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.173468] env[63021]: DEBUG nova.network.neutron [req-fa8c7c68-2a56-4cb3-b572-64272440557b req-2157dd1b-54d9-4d7e-bc03-e279b7c9d867 service nova] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Refreshing network info cache for port 8b1b2b2c-fec7-4c23-8349-bc89d9cc5b36 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 614.184944] env[63021]: DEBUG nova.network.neutron [-] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 614.339082] env[63021]: DEBUG nova.network.neutron [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Successfully created port: c25b24ca-5c74-4626-9705-e6c28e49a8c5 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 614.542804] env[63021]: DEBUG nova.compute.manager [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 614.687261] env[63021]: DEBUG nova.network.neutron [-] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.712719] env[63021]: DEBUG nova.network.neutron [req-fa8c7c68-2a56-4cb3-b572-64272440557b req-2157dd1b-54d9-4d7e-bc03-e279b7c9d867 service nova] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 614.810757] env[63021]: DEBUG nova.network.neutron [req-fa8c7c68-2a56-4cb3-b572-64272440557b req-2157dd1b-54d9-4d7e-bc03-e279b7c9d867 service nova] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.977673] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5a72720-9953-4f6b-b40f-7b4bdd5d0f89 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.986337] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-605b16a1-1ccc-474b-8291-39c1f274630d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.016980] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3f8f572-275f-43dc-be27-7bb0f891137a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.024503] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e8b2592-c71d-4339-92a0-fb6222c24bd8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.039092] env[63021]: DEBUG nova.compute.provider_tree [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 615.192191] env[63021]: INFO nova.compute.manager [-] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Took 1.03 seconds to deallocate network for instance. [ 615.193134] env[63021]: DEBUG nova.compute.claims [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 615.193134] env[63021]: DEBUG oslo_concurrency.lockutils [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.315185] env[63021]: DEBUG oslo_concurrency.lockutils [req-fa8c7c68-2a56-4cb3-b572-64272440557b req-2157dd1b-54d9-4d7e-bc03-e279b7c9d867 service nova] Releasing lock "refresh_cache-efd0e375-3134-4984-8f82-48b53c65e26c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.315185] env[63021]: DEBUG nova.compute.manager [req-fa8c7c68-2a56-4cb3-b572-64272440557b req-2157dd1b-54d9-4d7e-bc03-e279b7c9d867 service nova] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Received event network-vif-deleted-8b1b2b2c-fec7-4c23-8349-bc89d9cc5b36 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 615.361074] env[63021]: ERROR nova.compute.manager [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c25b24ca-5c74-4626-9705-e6c28e49a8c5, please check neutron logs for more information. [ 615.361074] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 615.361074] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 615.361074] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 615.361074] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 615.361074] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 615.361074] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 615.361074] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 615.361074] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.361074] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 615.361074] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.361074] env[63021]: ERROR nova.compute.manager raise self.value [ 615.361074] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 615.361074] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 615.361074] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.361074] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 615.361726] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.361726] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 615.361726] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c25b24ca-5c74-4626-9705-e6c28e49a8c5, please check neutron logs for more information. [ 615.361726] env[63021]: ERROR nova.compute.manager [ 615.361726] env[63021]: Traceback (most recent call last): [ 615.361726] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 615.361726] env[63021]: listener.cb(fileno) [ 615.361726] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 615.361726] env[63021]: result = function(*args, **kwargs) [ 615.361726] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 615.361726] env[63021]: return func(*args, **kwargs) [ 615.361726] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 615.361726] env[63021]: raise e [ 615.361726] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 615.361726] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 615.361726] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 615.361726] env[63021]: created_port_ids = self._update_ports_for_instance( [ 615.361726] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 615.361726] env[63021]: with excutils.save_and_reraise_exception(): [ 615.361726] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.361726] env[63021]: self.force_reraise() [ 615.361726] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.361726] env[63021]: raise self.value [ 615.361726] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 615.361726] env[63021]: updated_port = self._update_port( [ 615.361726] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.361726] env[63021]: _ensure_no_port_binding_failure(port) [ 615.361726] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.361726] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 615.362825] env[63021]: nova.exception.PortBindingFailed: Binding failed for port c25b24ca-5c74-4626-9705-e6c28e49a8c5, please check neutron logs for more information. [ 615.362825] env[63021]: Removing descriptor: 15 [ 615.542438] env[63021]: DEBUG nova.scheduler.client.report [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 615.552250] env[63021]: DEBUG nova.compute.manager [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 615.582568] env[63021]: DEBUG nova.virt.hardware [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 615.582810] env[63021]: DEBUG nova.virt.hardware [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 615.582964] env[63021]: DEBUG nova.virt.hardware [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 615.583156] env[63021]: DEBUG nova.virt.hardware [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 615.583303] env[63021]: DEBUG nova.virt.hardware [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 615.583474] env[63021]: DEBUG nova.virt.hardware [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 615.583684] env[63021]: DEBUG nova.virt.hardware [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 615.583836] env[63021]: DEBUG nova.virt.hardware [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 615.583998] env[63021]: DEBUG nova.virt.hardware [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 615.584190] env[63021]: DEBUG nova.virt.hardware [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 615.584376] env[63021]: DEBUG nova.virt.hardware [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 615.585203] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c5f863b-55c3-473a-9838-6b53ef1c2a87 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.594746] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db1d6033-5bdd-46fe-9fc9-ac31db9a958f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.609962] env[63021]: ERROR nova.compute.manager [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c25b24ca-5c74-4626-9705-e6c28e49a8c5, please check neutron logs for more information. [ 615.609962] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Traceback (most recent call last): [ 615.609962] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 615.609962] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] yield resources [ 615.609962] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 615.609962] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] self.driver.spawn(context, instance, image_meta, [ 615.609962] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 615.609962] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 615.609962] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 615.609962] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] vm_ref = self.build_virtual_machine(instance, [ 615.609962] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 615.610606] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] vif_infos = vmwarevif.get_vif_info(self._session, [ 615.610606] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 615.610606] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] for vif in network_info: [ 615.610606] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 615.610606] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] return self._sync_wrapper(fn, *args, **kwargs) [ 615.610606] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 615.610606] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] self.wait() [ 615.610606] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 615.610606] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] self[:] = self._gt.wait() [ 615.610606] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 615.610606] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] return self._exit_event.wait() [ 615.610606] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 615.610606] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] current.throw(*self._exc) [ 615.611277] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 615.611277] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] result = function(*args, **kwargs) [ 615.611277] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 615.611277] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] return func(*args, **kwargs) [ 615.611277] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 615.611277] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] raise e [ 615.611277] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 615.611277] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] nwinfo = self.network_api.allocate_for_instance( [ 615.611277] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 615.611277] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] created_port_ids = self._update_ports_for_instance( [ 615.611277] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 615.611277] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] with excutils.save_and_reraise_exception(): [ 615.611277] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.611970] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] self.force_reraise() [ 615.611970] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.611970] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] raise self.value [ 615.611970] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 615.611970] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] updated_port = self._update_port( [ 615.611970] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.611970] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] _ensure_no_port_binding_failure(port) [ 615.611970] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.611970] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] raise exception.PortBindingFailed(port_id=port['id']) [ 615.611970] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] nova.exception.PortBindingFailed: Binding failed for port c25b24ca-5c74-4626-9705-e6c28e49a8c5, please check neutron logs for more information. [ 615.611970] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] [ 615.611970] env[63021]: INFO nova.compute.manager [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Terminating instance [ 615.612471] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Acquiring lock "refresh_cache-92280cdf-bad2-4893-9a37-706cb1bcbac4" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.612606] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Acquired lock "refresh_cache-92280cdf-bad2-4893-9a37-706cb1bcbac4" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.612766] env[63021]: DEBUG nova.network.neutron [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 616.052697] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.525s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.053255] env[63021]: DEBUG nova.compute.manager [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 616.056139] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.819s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.056353] env[63021]: DEBUG nova.objects.instance [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Lazy-loading 'resources' on Instance uuid 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 616.137577] env[63021]: DEBUG nova.network.neutron [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 616.223903] env[63021]: DEBUG nova.network.neutron [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.257308] env[63021]: DEBUG nova.compute.manager [req-6e0f2c0b-9535-43b9-8bcf-c4d4fd255aa5 req-0f73d1da-f73f-448b-8480-1a1c279c0045 service nova] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Received event network-changed-c25b24ca-5c74-4626-9705-e6c28e49a8c5 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 616.257410] env[63021]: DEBUG nova.compute.manager [req-6e0f2c0b-9535-43b9-8bcf-c4d4fd255aa5 req-0f73d1da-f73f-448b-8480-1a1c279c0045 service nova] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Refreshing instance network info cache due to event network-changed-c25b24ca-5c74-4626-9705-e6c28e49a8c5. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 616.257579] env[63021]: DEBUG oslo_concurrency.lockutils [req-6e0f2c0b-9535-43b9-8bcf-c4d4fd255aa5 req-0f73d1da-f73f-448b-8480-1a1c279c0045 service nova] Acquiring lock "refresh_cache-92280cdf-bad2-4893-9a37-706cb1bcbac4" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.559400] env[63021]: DEBUG nova.compute.utils [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 616.563982] env[63021]: DEBUG nova.compute.manager [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 616.564280] env[63021]: DEBUG nova.network.neutron [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 616.617996] env[63021]: DEBUG nova.policy [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2aedf7c17a8e4a4caf7285f183f4f731', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '87484a3f9cf94549a4290f436544506f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 616.724973] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Releasing lock "refresh_cache-92280cdf-bad2-4893-9a37-706cb1bcbac4" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.725091] env[63021]: DEBUG nova.compute.manager [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 616.725282] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 616.725587] env[63021]: DEBUG oslo_concurrency.lockutils [req-6e0f2c0b-9535-43b9-8bcf-c4d4fd255aa5 req-0f73d1da-f73f-448b-8480-1a1c279c0045 service nova] Acquired lock "refresh_cache-92280cdf-bad2-4893-9a37-706cb1bcbac4" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.725866] env[63021]: DEBUG nova.network.neutron [req-6e0f2c0b-9535-43b9-8bcf-c4d4fd255aa5 req-0f73d1da-f73f-448b-8480-1a1c279c0045 service nova] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Refreshing network info cache for port c25b24ca-5c74-4626-9705-e6c28e49a8c5 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 616.727605] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b3b5b61a-3ad7-450c-bbfb-ee29b81a1440 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.745028] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a8a91bf-e094-485e-96a3-6a8b73efc528 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.772805] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 92280cdf-bad2-4893-9a37-706cb1bcbac4 could not be found. [ 616.773042] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 616.773240] env[63021]: INFO nova.compute.manager [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Took 0.05 seconds to destroy the instance on the hypervisor. [ 616.773486] env[63021]: DEBUG oslo.service.loopingcall [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 616.777251] env[63021]: DEBUG nova.compute.manager [-] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 616.777351] env[63021]: DEBUG nova.network.neutron [-] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 616.796736] env[63021]: DEBUG nova.network.neutron [-] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 616.912059] env[63021]: DEBUG nova.network.neutron [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Successfully created port: 758c2831-cb51-46c3-a612-cbbd0215f383 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 617.064685] env[63021]: DEBUG nova.compute.manager [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 617.074088] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d3bd757-eebb-41f3-8d49-8491b6422e44 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.083741] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7b8a3f8-efc7-4f78-b7f9-5b7e3074212f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.119038] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca281bf8-c8ee-43b5-bbb9-3ee5d0a655c3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.127247] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28b41de5-4a6c-403b-8a63-1ccbc3e02f93 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.141717] env[63021]: DEBUG nova.compute.provider_tree [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 617.219239] env[63021]: DEBUG nova.network.neutron [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Successfully created port: 607d2787-c3b2-4f2d-9449-88482ef3a334 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 617.252435] env[63021]: DEBUG nova.network.neutron [req-6e0f2c0b-9535-43b9-8bcf-c4d4fd255aa5 req-0f73d1da-f73f-448b-8480-1a1c279c0045 service nova] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 617.299737] env[63021]: DEBUG nova.network.neutron [-] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.480844] env[63021]: DEBUG nova.network.neutron [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Successfully created port: 99205906-07a6-4336-b4a9-8767c186cf52 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 617.532539] env[63021]: DEBUG nova.network.neutron [req-6e0f2c0b-9535-43b9-8bcf-c4d4fd255aa5 req-0f73d1da-f73f-448b-8480-1a1c279c0045 service nova] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.644286] env[63021]: DEBUG nova.scheduler.client.report [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 617.801274] env[63021]: INFO nova.compute.manager [-] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Took 1.02 seconds to deallocate network for instance. [ 617.806940] env[63021]: DEBUG nova.compute.claims [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 617.809019] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.037215] env[63021]: DEBUG oslo_concurrency.lockutils [req-6e0f2c0b-9535-43b9-8bcf-c4d4fd255aa5 req-0f73d1da-f73f-448b-8480-1a1c279c0045 service nova] Releasing lock "refresh_cache-92280cdf-bad2-4893-9a37-706cb1bcbac4" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.037539] env[63021]: DEBUG nova.compute.manager [req-6e0f2c0b-9535-43b9-8bcf-c4d4fd255aa5 req-0f73d1da-f73f-448b-8480-1a1c279c0045 service nova] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Received event network-vif-deleted-c25b24ca-5c74-4626-9705-e6c28e49a8c5 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 618.073992] env[63021]: DEBUG nova.compute.manager [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 618.099442] env[63021]: DEBUG nova.virt.hardware [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 618.099691] env[63021]: DEBUG nova.virt.hardware [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 618.099841] env[63021]: DEBUG nova.virt.hardware [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 618.100034] env[63021]: DEBUG nova.virt.hardware [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 618.100183] env[63021]: DEBUG nova.virt.hardware [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 618.100331] env[63021]: DEBUG nova.virt.hardware [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 618.100532] env[63021]: DEBUG nova.virt.hardware [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 618.100684] env[63021]: DEBUG nova.virt.hardware [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 618.100844] env[63021]: DEBUG nova.virt.hardware [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 618.100998] env[63021]: DEBUG nova.virt.hardware [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 618.101207] env[63021]: DEBUG nova.virt.hardware [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 618.102122] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05d35f20-4792-49b3-a921-ec9680597242 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.110437] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-129af55c-6ea8-48cd-bc53-a36f1706059d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.153519] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.095s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.154319] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.283s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.176088] env[63021]: INFO nova.scheduler.client.report [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Deleted allocations for instance 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd [ 618.430187] env[63021]: ERROR nova.compute.manager [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 758c2831-cb51-46c3-a612-cbbd0215f383, please check neutron logs for more information. [ 618.430187] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 618.430187] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.430187] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 618.430187] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.430187] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 618.430187] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.430187] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 618.430187] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.430187] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 618.430187] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.430187] env[63021]: ERROR nova.compute.manager raise self.value [ 618.430187] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.430187] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 618.430187] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.430187] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 618.430779] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.430779] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 618.430779] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 758c2831-cb51-46c3-a612-cbbd0215f383, please check neutron logs for more information. [ 618.430779] env[63021]: ERROR nova.compute.manager [ 618.430779] env[63021]: Traceback (most recent call last): [ 618.430779] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 618.430779] env[63021]: listener.cb(fileno) [ 618.430779] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.430779] env[63021]: result = function(*args, **kwargs) [ 618.430779] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 618.430779] env[63021]: return func(*args, **kwargs) [ 618.430779] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 618.430779] env[63021]: raise e [ 618.430779] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.430779] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 618.430779] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.430779] env[63021]: created_port_ids = self._update_ports_for_instance( [ 618.430779] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.430779] env[63021]: with excutils.save_and_reraise_exception(): [ 618.430779] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.430779] env[63021]: self.force_reraise() [ 618.430779] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.430779] env[63021]: raise self.value [ 618.430779] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.430779] env[63021]: updated_port = self._update_port( [ 618.430779] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.430779] env[63021]: _ensure_no_port_binding_failure(port) [ 618.430779] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.430779] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 618.431755] env[63021]: nova.exception.PortBindingFailed: Binding failed for port 758c2831-cb51-46c3-a612-cbbd0215f383, please check neutron logs for more information. [ 618.431755] env[63021]: Removing descriptor: 15 [ 618.431755] env[63021]: ERROR nova.compute.manager [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 758c2831-cb51-46c3-a612-cbbd0215f383, please check neutron logs for more information. [ 618.431755] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Traceback (most recent call last): [ 618.431755] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 618.431755] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] yield resources [ 618.431755] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 618.431755] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] self.driver.spawn(context, instance, image_meta, [ 618.431755] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 618.431755] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 618.431755] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 618.431755] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] vm_ref = self.build_virtual_machine(instance, [ 618.432258] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 618.432258] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] vif_infos = vmwarevif.get_vif_info(self._session, [ 618.432258] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 618.432258] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] for vif in network_info: [ 618.432258] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 618.432258] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] return self._sync_wrapper(fn, *args, **kwargs) [ 618.432258] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 618.432258] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] self.wait() [ 618.432258] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 618.432258] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] self[:] = self._gt.wait() [ 618.432258] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 618.432258] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] return self._exit_event.wait() [ 618.432258] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 618.432693] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] result = hub.switch() [ 618.432693] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 618.432693] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] return self.greenlet.switch() [ 618.432693] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.432693] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] result = function(*args, **kwargs) [ 618.432693] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 618.432693] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] return func(*args, **kwargs) [ 618.432693] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 618.432693] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] raise e [ 618.432693] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.432693] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] nwinfo = self.network_api.allocate_for_instance( [ 618.432693] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.432693] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] created_port_ids = self._update_ports_for_instance( [ 618.433133] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.433133] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] with excutils.save_and_reraise_exception(): [ 618.433133] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.433133] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] self.force_reraise() [ 618.433133] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.433133] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] raise self.value [ 618.433133] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.433133] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] updated_port = self._update_port( [ 618.433133] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.433133] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] _ensure_no_port_binding_failure(port) [ 618.433133] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.433133] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] raise exception.PortBindingFailed(port_id=port['id']) [ 618.433540] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] nova.exception.PortBindingFailed: Binding failed for port 758c2831-cb51-46c3-a612-cbbd0215f383, please check neutron logs for more information. [ 618.433540] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] [ 618.433540] env[63021]: INFO nova.compute.manager [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Terminating instance [ 618.435609] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Acquiring lock "refresh_cache-cee88fa5-c98e-405f-b41c-62d6661b983c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.435609] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Acquired lock "refresh_cache-cee88fa5-c98e-405f-b41c-62d6661b983c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.435609] env[63021]: DEBUG nova.network.neutron [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 618.468348] env[63021]: DEBUG nova.compute.manager [req-e48b5cb9-5eae-4f9e-95d9-93f4750a83ea req-8a8697d6-07f3-4c6c-bd46-405e5f0a3e5b service nova] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Received event network-changed-758c2831-cb51-46c3-a612-cbbd0215f383 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 618.468594] env[63021]: DEBUG nova.compute.manager [req-e48b5cb9-5eae-4f9e-95d9-93f4750a83ea req-8a8697d6-07f3-4c6c-bd46-405e5f0a3e5b service nova] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Refreshing instance network info cache due to event network-changed-758c2831-cb51-46c3-a612-cbbd0215f383. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 618.468778] env[63021]: DEBUG oslo_concurrency.lockutils [req-e48b5cb9-5eae-4f9e-95d9-93f4750a83ea req-8a8697d6-07f3-4c6c-bd46-405e5f0a3e5b service nova] Acquiring lock "refresh_cache-cee88fa5-c98e-405f-b41c-62d6661b983c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.475300] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Acquiring lock "c2523d70-12dd-444b-8839-b97378ef687e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.475402] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Lock "c2523d70-12dd-444b-8839-b97378ef687e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.685614] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b5b54dca-de7b-48b3-9464-54a03f84d3e9 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Lock "5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.167s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.997394] env[63021]: DEBUG nova.network.neutron [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 619.031636] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Acquiring lock "0950cbd3-0708-4994-8d2d-053aecc06cdf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.031886] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Lock "0950cbd3-0708-4994-8d2d-053aecc06cdf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.032129] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Acquiring lock "0950cbd3-0708-4994-8d2d-053aecc06cdf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.032363] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Lock "0950cbd3-0708-4994-8d2d-053aecc06cdf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.032641] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Lock "0950cbd3-0708-4994-8d2d-053aecc06cdf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.038914] env[63021]: INFO nova.compute.manager [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Terminating instance [ 619.042351] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Acquiring lock "refresh_cache-0950cbd3-0708-4994-8d2d-053aecc06cdf" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.042507] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Acquired lock "refresh_cache-0950cbd3-0708-4994-8d2d-053aecc06cdf" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.042672] env[63021]: DEBUG nova.network.neutron [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 619.086030] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24e0c5da-74b1-44c5-ae1b-313b8a835675 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.094564] env[63021]: DEBUG nova.network.neutron [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.097531] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9b4d21a-0e68-4b5d-b53b-d593c5782e13 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.129604] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b563ebe-5477-4d68-a73a-ba8e6750a197 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.139062] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ed577dc-f229-4986-a918-f16f9140c74c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.153419] env[63021]: DEBUG nova.compute.provider_tree [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 619.560896] env[63021]: DEBUG nova.network.neutron [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 619.603168] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Releasing lock "refresh_cache-cee88fa5-c98e-405f-b41c-62d6661b983c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.603264] env[63021]: DEBUG nova.compute.manager [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 619.603374] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 619.603676] env[63021]: DEBUG oslo_concurrency.lockutils [req-e48b5cb9-5eae-4f9e-95d9-93f4750a83ea req-8a8697d6-07f3-4c6c-bd46-405e5f0a3e5b service nova] Acquired lock "refresh_cache-cee88fa5-c98e-405f-b41c-62d6661b983c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.603867] env[63021]: DEBUG nova.network.neutron [req-e48b5cb9-5eae-4f9e-95d9-93f4750a83ea req-8a8697d6-07f3-4c6c-bd46-405e5f0a3e5b service nova] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Refreshing network info cache for port 758c2831-cb51-46c3-a612-cbbd0215f383 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 619.605037] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-be5fb3fe-4bcf-47a0-bec7-79d6d02ae9bb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.607839] env[63021]: DEBUG nova.network.neutron [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.616602] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfce5f74-3cfe-43d3-af67-91bd6daba49d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.640439] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance cee88fa5-c98e-405f-b41c-62d6661b983c could not be found. [ 619.640439] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 619.640439] env[63021]: INFO nova.compute.manager [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 619.640439] env[63021]: DEBUG oslo.service.loopingcall [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 619.640439] env[63021]: DEBUG nova.compute.manager [-] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 619.640439] env[63021]: DEBUG nova.network.neutron [-] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 619.656240] env[63021]: DEBUG nova.scheduler.client.report [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 619.702859] env[63021]: DEBUG nova.network.neutron [-] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.113649] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Releasing lock "refresh_cache-0950cbd3-0708-4994-8d2d-053aecc06cdf" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.114056] env[63021]: DEBUG nova.compute.manager [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 620.114249] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 620.115290] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9e030d0-b081-4e5e-8cda-01fa7e9dd413 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.123783] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 620.124172] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b06c2231-766d-412b-b13a-88bcc86bfbf7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.132326] env[63021]: DEBUG oslo_vmware.api [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 620.132326] env[63021]: value = "task-1293576" [ 620.132326] env[63021]: _type = "Task" [ 620.132326] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.140875] env[63021]: DEBUG oslo_vmware.api [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293576, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.160972] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.007s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.161693] env[63021]: ERROR nova.compute.manager [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9067a00a-e182-455c-88cb-1e0d7666dfe5, please check neutron logs for more information. [ 620.161693] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Traceback (most recent call last): [ 620.161693] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 620.161693] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] self.driver.spawn(context, instance, image_meta, [ 620.161693] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 620.161693] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 620.161693] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 620.161693] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] vm_ref = self.build_virtual_machine(instance, [ 620.161693] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 620.161693] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] vif_infos = vmwarevif.get_vif_info(self._session, [ 620.161693] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 620.161942] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] for vif in network_info: [ 620.161942] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 620.161942] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] return self._sync_wrapper(fn, *args, **kwargs) [ 620.161942] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 620.161942] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] self.wait() [ 620.161942] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 620.161942] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] self[:] = self._gt.wait() [ 620.161942] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 620.161942] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] return self._exit_event.wait() [ 620.161942] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 620.161942] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] result = hub.switch() [ 620.161942] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 620.161942] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] return self.greenlet.switch() [ 620.162216] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 620.162216] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] result = function(*args, **kwargs) [ 620.162216] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 620.162216] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] return func(*args, **kwargs) [ 620.162216] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 620.162216] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] raise e [ 620.162216] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 620.162216] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] nwinfo = self.network_api.allocate_for_instance( [ 620.162216] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 620.162216] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] created_port_ids = self._update_ports_for_instance( [ 620.162216] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 620.162216] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] with excutils.save_and_reraise_exception(): [ 620.162216] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.162461] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] self.force_reraise() [ 620.162461] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.162461] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] raise self.value [ 620.162461] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 620.162461] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] updated_port = self._update_port( [ 620.162461] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.162461] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] _ensure_no_port_binding_failure(port) [ 620.162461] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.162461] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] raise exception.PortBindingFailed(port_id=port['id']) [ 620.162461] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] nova.exception.PortBindingFailed: Binding failed for port 9067a00a-e182-455c-88cb-1e0d7666dfe5, please check neutron logs for more information. [ 620.162461] env[63021]: ERROR nova.compute.manager [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] [ 620.162737] env[63021]: DEBUG nova.compute.utils [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Binding failed for port 9067a00a-e182-455c-88cb-1e0d7666dfe5, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 620.165411] env[63021]: DEBUG nova.network.neutron [req-e48b5cb9-5eae-4f9e-95d9-93f4750a83ea req-8a8697d6-07f3-4c6c-bd46-405e5f0a3e5b service nova] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.167598] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.577s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 620.175787] env[63021]: DEBUG nova.compute.manager [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Build of instance 082e6e14-c5f7-4ca4-b739-9d997a77b8f9 was re-scheduled: Binding failed for port 9067a00a-e182-455c-88cb-1e0d7666dfe5, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 620.175787] env[63021]: DEBUG nova.compute.manager [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 620.175787] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Acquiring lock "refresh_cache-082e6e14-c5f7-4ca4-b739-9d997a77b8f9" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 620.175787] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Acquired lock "refresh_cache-082e6e14-c5f7-4ca4-b739-9d997a77b8f9" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.176135] env[63021]: DEBUG nova.network.neutron [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 620.302548] env[63021]: DEBUG nova.network.neutron [req-e48b5cb9-5eae-4f9e-95d9-93f4750a83ea req-8a8697d6-07f3-4c6c-bd46-405e5f0a3e5b service nova] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.520635] env[63021]: DEBUG nova.compute.manager [req-a5998cc9-42bb-421e-b0a3-cf3db3aeeab8 req-9d0f4fdf-3db4-4380-bbdf-3a1551fb9237 service nova] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Received event network-vif-deleted-758c2831-cb51-46c3-a612-cbbd0215f383 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 620.642021] env[63021]: DEBUG oslo_vmware.api [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293576, 'name': PowerOffVM_Task, 'duration_secs': 0.126943} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.642340] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 620.642550] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 620.642747] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3fbe38cb-a85b-46dd-a5e3-9738069b4d22 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.666055] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 620.666268] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 620.666454] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Deleting the datastore file [datastore1] 0950cbd3-0708-4994-8d2d-053aecc06cdf {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 620.666697] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7ba89250-5bd7-4168-b85d-f268edc765a5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.672704] env[63021]: DEBUG oslo_vmware.api [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for the task: (returnval){ [ 620.672704] env[63021]: value = "task-1293578" [ 620.672704] env[63021]: _type = "Task" [ 620.672704] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.686042] env[63021]: DEBUG oslo_vmware.api [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293578, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.710962] env[63021]: DEBUG nova.network.neutron [-] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.714101] env[63021]: DEBUG nova.network.neutron [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.794218] env[63021]: DEBUG nova.network.neutron [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.805637] env[63021]: DEBUG oslo_concurrency.lockutils [req-e48b5cb9-5eae-4f9e-95d9-93f4750a83ea req-8a8697d6-07f3-4c6c-bd46-405e5f0a3e5b service nova] Releasing lock "refresh_cache-cee88fa5-c98e-405f-b41c-62d6661b983c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.994558] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-027e5a1f-8592-4b46-88be-451ff9238521 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.001812] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd60d2f7-effd-4cd0-9b35-3208485715d0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.032316] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c72a02d1-a220-48d3-a24a-9c540c0dab22 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.039443] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d3ffd76-3637-49d4-b1d8-606d8b28fbff {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.053547] env[63021]: DEBUG nova.compute.provider_tree [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 621.182440] env[63021]: DEBUG oslo_vmware.api [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Task: {'id': task-1293578, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.109023} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.182755] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 621.182924] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 621.183144] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 621.183350] env[63021]: INFO nova.compute.manager [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Took 1.07 seconds to destroy the instance on the hypervisor. [ 621.183607] env[63021]: DEBUG oslo.service.loopingcall [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 621.183838] env[63021]: DEBUG nova.compute.manager [-] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 621.183952] env[63021]: DEBUG nova.network.neutron [-] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 621.196332] env[63021]: DEBUG nova.network.neutron [-] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.213136] env[63021]: INFO nova.compute.manager [-] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Took 1.57 seconds to deallocate network for instance. [ 621.215377] env[63021]: DEBUG nova.compute.claims [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 621.215550] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.296486] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Releasing lock "refresh_cache-082e6e14-c5f7-4ca4-b739-9d997a77b8f9" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 621.296724] env[63021]: DEBUG nova.compute.manager [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 621.296906] env[63021]: DEBUG nova.compute.manager [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 621.297083] env[63021]: DEBUG nova.network.neutron [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 621.314732] env[63021]: DEBUG nova.network.neutron [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.557036] env[63021]: DEBUG nova.scheduler.client.report [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 621.700650] env[63021]: DEBUG nova.network.neutron [-] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.817641] env[63021]: DEBUG nova.network.neutron [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.061728] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.894s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.062428] env[63021]: ERROR nova.compute.manager [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 95e81d9c-71a8-4616-a6a9-1cc0eeef6164, please check neutron logs for more information. [ 622.062428] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Traceback (most recent call last): [ 622.062428] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 622.062428] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] self.driver.spawn(context, instance, image_meta, [ 622.062428] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 622.062428] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 622.062428] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 622.062428] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] vm_ref = self.build_virtual_machine(instance, [ 622.062428] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 622.062428] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] vif_infos = vmwarevif.get_vif_info(self._session, [ 622.062428] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 622.062700] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] for vif in network_info: [ 622.062700] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 622.062700] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] return self._sync_wrapper(fn, *args, **kwargs) [ 622.062700] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 622.062700] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] self.wait() [ 622.062700] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 622.062700] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] self[:] = self._gt.wait() [ 622.062700] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 622.062700] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] return self._exit_event.wait() [ 622.062700] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 622.062700] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] result = hub.switch() [ 622.062700] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 622.062700] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] return self.greenlet.switch() [ 622.062957] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 622.062957] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] result = function(*args, **kwargs) [ 622.062957] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 622.062957] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] return func(*args, **kwargs) [ 622.062957] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 622.062957] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] raise e [ 622.062957] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 622.062957] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] nwinfo = self.network_api.allocate_for_instance( [ 622.062957] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 622.062957] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] created_port_ids = self._update_ports_for_instance( [ 622.062957] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 622.062957] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] with excutils.save_and_reraise_exception(): [ 622.062957] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.063281] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] self.force_reraise() [ 622.063281] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.063281] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] raise self.value [ 622.063281] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 622.063281] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] updated_port = self._update_port( [ 622.063281] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.063281] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] _ensure_no_port_binding_failure(port) [ 622.063281] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.063281] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] raise exception.PortBindingFailed(port_id=port['id']) [ 622.063281] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] nova.exception.PortBindingFailed: Binding failed for port 95e81d9c-71a8-4616-a6a9-1cc0eeef6164, please check neutron logs for more information. [ 622.063281] env[63021]: ERROR nova.compute.manager [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] [ 622.063550] env[63021]: DEBUG nova.compute.utils [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Binding failed for port 95e81d9c-71a8-4616-a6a9-1cc0eeef6164, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 622.064661] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.762s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.067950] env[63021]: DEBUG nova.compute.manager [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Build of instance 9f64a5ad-7852-4772-9b8c-2ec1537204f3 was re-scheduled: Binding failed for port 95e81d9c-71a8-4616-a6a9-1cc0eeef6164, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 622.068411] env[63021]: DEBUG nova.compute.manager [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 622.068670] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Acquiring lock "refresh_cache-9f64a5ad-7852-4772-9b8c-2ec1537204f3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.068844] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Acquired lock "refresh_cache-9f64a5ad-7852-4772-9b8c-2ec1537204f3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.069102] env[63021]: DEBUG nova.network.neutron [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 622.202893] env[63021]: INFO nova.compute.manager [-] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Took 1.02 seconds to deallocate network for instance. [ 622.320231] env[63021]: INFO nova.compute.manager [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] [instance: 082e6e14-c5f7-4ca4-b739-9d997a77b8f9] Took 1.02 seconds to deallocate network for instance. [ 622.603849] env[63021]: DEBUG nova.network.neutron [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.708465] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.717813] env[63021]: DEBUG nova.network.neutron [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.912586] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-911f220c-0111-4571-9ba6-fa7a5895cc25 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.920381] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cc407ea-b7af-4a61-86ed-1259fc34308d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.950314] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5719a7e2-97a5-4b69-9e0b-05e42598d168 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.957636] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ed656da-e484-44ca-9514-14fb8c152856 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.970877] env[63021]: DEBUG nova.compute.provider_tree [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 623.221051] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Releasing lock "refresh_cache-9f64a5ad-7852-4772-9b8c-2ec1537204f3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.221051] env[63021]: DEBUG nova.compute.manager [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 623.221399] env[63021]: DEBUG nova.compute.manager [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 623.221399] env[63021]: DEBUG nova.network.neutron [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 623.251048] env[63021]: DEBUG nova.network.neutron [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 623.354024] env[63021]: INFO nova.scheduler.client.report [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Deleted allocations for instance 082e6e14-c5f7-4ca4-b739-9d997a77b8f9 [ 623.474922] env[63021]: DEBUG nova.scheduler.client.report [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 623.756820] env[63021]: DEBUG nova.network.neutron [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.861274] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2ce98153-ad60-4d53-bc1b-160357cc5339 tempest-ServersWithSpecificFlavorTestJSON-1324117861 tempest-ServersWithSpecificFlavorTestJSON-1324117861-project-member] Lock "082e6e14-c5f7-4ca4-b739-9d997a77b8f9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.056s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.980115] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.915s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.980778] env[63021]: ERROR nova.compute.manager [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 663a49a6-fb86-4199-b1f0-fbf8f32e806c, please check neutron logs for more information. [ 623.980778] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Traceback (most recent call last): [ 623.980778] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 623.980778] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] self.driver.spawn(context, instance, image_meta, [ 623.980778] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 623.980778] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] self._vmops.spawn(context, instance, image_meta, injected_files, [ 623.980778] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 623.980778] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] vm_ref = self.build_virtual_machine(instance, [ 623.980778] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 623.980778] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] vif_infos = vmwarevif.get_vif_info(self._session, [ 623.980778] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 623.981222] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] for vif in network_info: [ 623.981222] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 623.981222] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] return self._sync_wrapper(fn, *args, **kwargs) [ 623.981222] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 623.981222] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] self.wait() [ 623.981222] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 623.981222] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] self[:] = self._gt.wait() [ 623.981222] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 623.981222] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] return self._exit_event.wait() [ 623.981222] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 623.981222] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] result = hub.switch() [ 623.981222] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 623.981222] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] return self.greenlet.switch() [ 623.981547] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 623.981547] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] result = function(*args, **kwargs) [ 623.981547] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 623.981547] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] return func(*args, **kwargs) [ 623.981547] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 623.981547] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] raise e [ 623.981547] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 623.981547] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] nwinfo = self.network_api.allocate_for_instance( [ 623.981547] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 623.981547] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] created_port_ids = self._update_ports_for_instance( [ 623.981547] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 623.981547] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] with excutils.save_and_reraise_exception(): [ 623.981547] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.981871] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] self.force_reraise() [ 623.981871] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.981871] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] raise self.value [ 623.981871] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 623.981871] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] updated_port = self._update_port( [ 623.981871] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.981871] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] _ensure_no_port_binding_failure(port) [ 623.981871] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.981871] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] raise exception.PortBindingFailed(port_id=port['id']) [ 623.981871] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] nova.exception.PortBindingFailed: Binding failed for port 663a49a6-fb86-4199-b1f0-fbf8f32e806c, please check neutron logs for more information. [ 623.981871] env[63021]: ERROR nova.compute.manager [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] [ 623.982157] env[63021]: DEBUG nova.compute.utils [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Binding failed for port 663a49a6-fb86-4199-b1f0-fbf8f32e806c, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 623.982751] env[63021]: DEBUG oslo_concurrency.lockutils [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.125s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.987029] env[63021]: DEBUG nova.compute.manager [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Build of instance 1fa1c7c7-b9fe-4e60-8306-3c263c517c28 was re-scheduled: Binding failed for port 663a49a6-fb86-4199-b1f0-fbf8f32e806c, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 623.987029] env[63021]: DEBUG nova.compute.manager [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 623.987029] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Acquiring lock "refresh_cache-1fa1c7c7-b9fe-4e60-8306-3c263c517c28" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 623.987029] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Acquired lock "refresh_cache-1fa1c7c7-b9fe-4e60-8306-3c263c517c28" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.987222] env[63021]: DEBUG nova.network.neutron [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 624.259290] env[63021]: INFO nova.compute.manager [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] [instance: 9f64a5ad-7852-4772-9b8c-2ec1537204f3] Took 1.04 seconds to deallocate network for instance. [ 624.362688] env[63021]: DEBUG nova.compute.manager [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 624.514362] env[63021]: DEBUG nova.network.neutron [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 624.616300] env[63021]: DEBUG nova.network.neutron [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.884923] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.914903] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20043d74-768e-45d3-af1d-304c7c50f187 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.923100] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5785156e-1703-43c8-92a3-ddcd86435355 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.955282] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1e6def9-3d4f-4371-8c7f-33dab8db62a0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.964017] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31bc05f5-3c94-4e4d-9d01-e7d068d57ee9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.979246] env[63021]: DEBUG nova.compute.provider_tree [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 625.119170] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Releasing lock "refresh_cache-1fa1c7c7-b9fe-4e60-8306-3c263c517c28" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.119368] env[63021]: DEBUG nova.compute.manager [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 625.119553] env[63021]: DEBUG nova.compute.manager [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 625.119765] env[63021]: DEBUG nova.network.neutron [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 625.136259] env[63021]: DEBUG nova.network.neutron [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.328461] env[63021]: INFO nova.scheduler.client.report [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Deleted allocations for instance 9f64a5ad-7852-4772-9b8c-2ec1537204f3 [ 625.486026] env[63021]: DEBUG nova.scheduler.client.report [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 625.639520] env[63021]: DEBUG nova.network.neutron [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.837195] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2dee7d6a-8ed2-4f99-9885-e8ac570921e2 tempest-ServerActionsTestOtherA-88801427 tempest-ServerActionsTestOtherA-88801427-project-member] Lock "9f64a5ad-7852-4772-9b8c-2ec1537204f3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.023s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.991023] env[63021]: DEBUG oslo_concurrency.lockutils [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.006s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.991023] env[63021]: ERROR nova.compute.manager [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5522f00b-405a-4de6-9f40-35ad30faab5c, please check neutron logs for more information. [ 625.991023] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Traceback (most recent call last): [ 625.991023] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 625.991023] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] self.driver.spawn(context, instance, image_meta, [ 625.991023] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 625.991023] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 625.991023] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 625.991023] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] vm_ref = self.build_virtual_machine(instance, [ 625.991769] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 625.991769] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] vif_infos = vmwarevif.get_vif_info(self._session, [ 625.991769] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 625.991769] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] for vif in network_info: [ 625.991769] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 625.991769] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] return self._sync_wrapper(fn, *args, **kwargs) [ 625.991769] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 625.991769] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] self.wait() [ 625.991769] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 625.991769] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] self[:] = self._gt.wait() [ 625.991769] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 625.991769] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] return self._exit_event.wait() [ 625.991769] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 625.992190] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] result = hub.switch() [ 625.992190] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 625.992190] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] return self.greenlet.switch() [ 625.992190] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 625.992190] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] result = function(*args, **kwargs) [ 625.992190] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 625.992190] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] return func(*args, **kwargs) [ 625.992190] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 625.992190] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] raise e [ 625.992190] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 625.992190] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] nwinfo = self.network_api.allocate_for_instance( [ 625.992190] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 625.992190] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] created_port_ids = self._update_ports_for_instance( [ 625.992524] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 625.992524] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] with excutils.save_and_reraise_exception(): [ 625.992524] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.992524] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] self.force_reraise() [ 625.992524] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.992524] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] raise self.value [ 625.992524] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 625.992524] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] updated_port = self._update_port( [ 625.992524] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.992524] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] _ensure_no_port_binding_failure(port) [ 625.992524] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.992524] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] raise exception.PortBindingFailed(port_id=port['id']) [ 625.992865] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] nova.exception.PortBindingFailed: Binding failed for port 5522f00b-405a-4de6-9f40-35ad30faab5c, please check neutron logs for more information. [ 625.992865] env[63021]: ERROR nova.compute.manager [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] [ 625.992865] env[63021]: DEBUG nova.compute.utils [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Binding failed for port 5522f00b-405a-4de6-9f40-35ad30faab5c, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 625.997030] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.232s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.997030] env[63021]: DEBUG nova.objects.instance [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Lazy-loading 'resources' on Instance uuid 9d474bd0-e825-49f2-9715-14d85495718e {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 625.997030] env[63021]: DEBUG nova.compute.manager [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Build of instance ee9bdcd3-03ac-4093-bf21-877a2a8847a0 was re-scheduled: Binding failed for port 5522f00b-405a-4de6-9f40-35ad30faab5c, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 625.997030] env[63021]: DEBUG nova.compute.manager [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 625.997408] env[63021]: DEBUG oslo_concurrency.lockutils [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Acquiring lock "refresh_cache-ee9bdcd3-03ac-4093-bf21-877a2a8847a0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.997408] env[63021]: DEBUG oslo_concurrency.lockutils [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Acquired lock "refresh_cache-ee9bdcd3-03ac-4093-bf21-877a2a8847a0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.997408] env[63021]: DEBUG nova.network.neutron [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 626.141890] env[63021]: INFO nova.compute.manager [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] [instance: 1fa1c7c7-b9fe-4e60-8306-3c263c517c28] Took 1.02 seconds to deallocate network for instance. [ 626.344152] env[63021]: DEBUG nova.compute.manager [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 626.529818] env[63021]: DEBUG nova.network.neutron [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 626.882388] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.912498] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71dd66ad-5fac-491c-83bd-07b3b743f04a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.916018] env[63021]: DEBUG nova.network.neutron [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.924603] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e262f19-2129-419b-a1e0-f93d2bf41626 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.963230] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af92532e-932c-47f1-99d2-6ea9ce74ec0d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.972760] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01de9441-29eb-471e-a10d-ce1c379591eb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.987569] env[63021]: DEBUG nova.compute.provider_tree [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 627.180019] env[63021]: INFO nova.scheduler.client.report [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Deleted allocations for instance 1fa1c7c7-b9fe-4e60-8306-3c263c517c28 [ 627.424956] env[63021]: DEBUG oslo_concurrency.lockutils [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Releasing lock "refresh_cache-ee9bdcd3-03ac-4093-bf21-877a2a8847a0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.424956] env[63021]: DEBUG nova.compute.manager [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 627.424956] env[63021]: DEBUG nova.compute.manager [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 627.425241] env[63021]: DEBUG nova.network.neutron [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 627.445890] env[63021]: DEBUG nova.network.neutron [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 627.491869] env[63021]: DEBUG nova.scheduler.client.report [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 627.685578] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1121bb97-6883-4974-aebc-42b2ddeb383f tempest-VolumesAssistedSnapshotsTest-700683139 tempest-VolumesAssistedSnapshotsTest-700683139-project-member] Lock "1fa1c7c7-b9fe-4e60-8306-3c263c517c28" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.359s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.948709] env[63021]: DEBUG nova.network.neutron [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.006130] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.012s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 628.008933] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.029s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.012185] env[63021]: INFO nova.compute.claims [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 628.033092] env[63021]: INFO nova.scheduler.client.report [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Deleted allocations for instance 9d474bd0-e825-49f2-9715-14d85495718e [ 628.188139] env[63021]: DEBUG nova.compute.manager [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 628.452137] env[63021]: INFO nova.compute.manager [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] [instance: ee9bdcd3-03ac-4093-bf21-877a2a8847a0] Took 1.03 seconds to deallocate network for instance. [ 628.545577] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3eb8e12c-0f5c-48d6-9d55-bb7c3d029a14 tempest-ServerDiagnosticsV248Test-332692924 tempest-ServerDiagnosticsV248Test-332692924-project-member] Lock "9d474bd0-e825-49f2-9715-14d85495718e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.474s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 628.715561] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.382892] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5f062d3-235f-4f9b-b15b-26a9a29a0db4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.392138] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e1ce14c-f222-4aeb-86dd-cd1cca93d0a7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.423896] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-067f0b31-c983-4adc-8e4f-0d2c2b2e0142 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.434108] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fff5e09-594a-40d9-b2c8-074e0070adff {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.449809] env[63021]: DEBUG nova.compute.provider_tree [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 629.482105] env[63021]: INFO nova.scheduler.client.report [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Deleted allocations for instance ee9bdcd3-03ac-4093-bf21-877a2a8847a0 [ 629.954133] env[63021]: DEBUG nova.scheduler.client.report [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 629.995273] env[63021]: DEBUG oslo_concurrency.lockutils [None req-328441a9-d75b-4998-a2a3-62b21a308744 tempest-AttachInterfacesV270Test-651199283 tempest-AttachInterfacesV270Test-651199283-project-member] Lock "ee9bdcd3-03ac-4093-bf21-877a2a8847a0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.618s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.008055] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 630.008945] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 630.459976] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.451s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.460546] env[63021]: DEBUG nova.compute.manager [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 630.466215] env[63021]: DEBUG oslo_concurrency.lockutils [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.703s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.499490] env[63021]: DEBUG nova.compute.manager [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 630.520298] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 630.521117] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Starting heal instance info cache {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 630.521280] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Rebuilding the list of instances to heal {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 630.969866] env[63021]: DEBUG nova.compute.utils [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 630.975544] env[63021]: DEBUG nova.compute.manager [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 630.975727] env[63021]: DEBUG nova.network.neutron [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 631.021682] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.027786] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Skipping network cache update for instance because it is Building. {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 631.027786] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Skipping network cache update for instance because it is Building. {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 631.027786] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Skipping network cache update for instance because it is Building. {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 631.027786] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Skipping network cache update for instance because it is Building. {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 631.027786] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Skipping network cache update for instance because it is Building. {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 631.048101] env[63021]: DEBUG nova.policy [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4b55e7e8f5864da58b9327ed79f1bf6a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fdec2707371345d49fb59030cab5d6b0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 631.066132] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquiring lock "refresh_cache-0950cbd3-0708-4994-8d2d-053aecc06cdf" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.066132] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquired lock "refresh_cache-0950cbd3-0708-4994-8d2d-053aecc06cdf" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.066132] env[63021]: DEBUG nova.network.neutron [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Forcefully refreshing network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 631.066132] env[63021]: DEBUG nova.objects.instance [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lazy-loading 'info_cache' on Instance uuid 0950cbd3-0708-4994-8d2d-053aecc06cdf {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 631.357309] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26e8c441-0c07-456d-b46f-ecba0605429c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.365249] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb214f81-07f3-4944-bdf1-0e21c619cad2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.398835] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40cb2668-8326-47f1-a4d9-df2e7db24b2d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.408125] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1170ff3a-c974-4a9d-ac1f-b70c86e89850 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.423302] env[63021]: DEBUG nova.compute.provider_tree [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 631.478986] env[63021]: DEBUG nova.compute.manager [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 631.530224] env[63021]: DEBUG nova.network.neutron [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Successfully created port: af20fd90-4fdd-4898-9acf-ee54c4d9f27b {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 632.167466] env[63021]: DEBUG nova.scheduler.client.report [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 632.251089] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Acquiring lock "26896531-07eb-4e44-8dd8-e5237bfd7c39" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.251629] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Lock "26896531-07eb-4e44-8dd8-e5237bfd7c39" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.678722] env[63021]: DEBUG oslo_concurrency.lockutils [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.215s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.679666] env[63021]: ERROR nova.compute.manager [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 36395c0e-99a7-4f1c-837f-5ed5718b3abc, please check neutron logs for more information. [ 632.679666] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Traceback (most recent call last): [ 632.679666] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 632.679666] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] self.driver.spawn(context, instance, image_meta, [ 632.679666] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 632.679666] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 632.679666] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 632.679666] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] vm_ref = self.build_virtual_machine(instance, [ 632.679666] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 632.679666] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] vif_infos = vmwarevif.get_vif_info(self._session, [ 632.679666] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 632.680015] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] for vif in network_info: [ 632.680015] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 632.680015] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] return self._sync_wrapper(fn, *args, **kwargs) [ 632.680015] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 632.680015] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] self.wait() [ 632.680015] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 632.680015] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] self[:] = self._gt.wait() [ 632.680015] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 632.680015] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] return self._exit_event.wait() [ 632.680015] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 632.680015] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] current.throw(*self._exc) [ 632.680015] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.680015] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] result = function(*args, **kwargs) [ 632.680302] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 632.680302] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] return func(*args, **kwargs) [ 632.680302] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 632.680302] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] raise e [ 632.680302] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.680302] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] nwinfo = self.network_api.allocate_for_instance( [ 632.680302] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 632.680302] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] created_port_ids = self._update_ports_for_instance( [ 632.680302] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 632.680302] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] with excutils.save_and_reraise_exception(): [ 632.680302] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.680302] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] self.force_reraise() [ 632.680302] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.681202] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] raise self.value [ 632.681202] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 632.681202] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] updated_port = self._update_port( [ 632.681202] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.681202] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] _ensure_no_port_binding_failure(port) [ 632.681202] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.681202] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] raise exception.PortBindingFailed(port_id=port['id']) [ 632.681202] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] nova.exception.PortBindingFailed: Binding failed for port 36395c0e-99a7-4f1c-837f-5ed5718b3abc, please check neutron logs for more information. [ 632.681202] env[63021]: ERROR nova.compute.manager [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] [ 632.683977] env[63021]: DEBUG nova.compute.utils [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Binding failed for port 36395c0e-99a7-4f1c-837f-5ed5718b3abc, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 632.684238] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.191s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.687639] env[63021]: DEBUG nova.compute.manager [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 632.691127] env[63021]: DEBUG nova.compute.manager [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Build of instance e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7 was re-scheduled: Binding failed for port 36395c0e-99a7-4f1c-837f-5ed5718b3abc, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 632.691127] env[63021]: DEBUG nova.compute.manager [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 632.691554] env[63021]: DEBUG oslo_concurrency.lockutils [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Acquiring lock "refresh_cache-e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 632.691554] env[63021]: DEBUG oslo_concurrency.lockutils [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Acquired lock "refresh_cache-e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.691685] env[63021]: DEBUG nova.network.neutron [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 632.720807] env[63021]: DEBUG nova.network.neutron [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.729281] env[63021]: DEBUG nova.virt.hardware [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 632.729281] env[63021]: DEBUG nova.virt.hardware [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 632.729281] env[63021]: DEBUG nova.virt.hardware [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 632.729426] env[63021]: DEBUG nova.virt.hardware [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 632.729426] env[63021]: DEBUG nova.virt.hardware [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 632.729426] env[63021]: DEBUG nova.virt.hardware [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 632.729509] env[63021]: DEBUG nova.virt.hardware [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 632.729612] env[63021]: DEBUG nova.virt.hardware [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 632.729777] env[63021]: DEBUG nova.virt.hardware [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 632.729933] env[63021]: DEBUG nova.virt.hardware [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 632.730226] env[63021]: DEBUG nova.virt.hardware [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 632.732463] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06c7cef1-8a13-4dfa-b65f-68a9ef1b3328 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.743603] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-531f4b20-d002-44fc-88e7-89f0a529cd8b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.213188] env[63021]: DEBUG nova.network.neutron [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.361030] env[63021]: DEBUG nova.network.neutron [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.371247] env[63021]: DEBUG nova.network.neutron [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.435615] env[63021]: DEBUG oslo_concurrency.lockutils [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Acquiring lock "274b001a-443f-463c-a657-652f64ba1de8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.435886] env[63021]: DEBUG oslo_concurrency.lockutils [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Lock "274b001a-443f-463c-a657-652f64ba1de8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.603172] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e423295e-141a-4fbe-8fda-c53e9fb9d65a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.610560] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20c01376-77e1-4a45-9f5e-38a632cfb67e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.641430] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a5a9f68-0041-43e5-8db3-7199e06553f0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.649592] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecfc4699-1e74-449b-8259-1fc5b2daab1f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.665272] env[63021]: DEBUG nova.compute.provider_tree [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 633.865365] env[63021]: DEBUG oslo_concurrency.lockutils [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Releasing lock "refresh_cache-e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.865701] env[63021]: DEBUG nova.compute.manager [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 633.865912] env[63021]: DEBUG nova.compute.manager [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 633.866100] env[63021]: DEBUG nova.network.neutron [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 633.871462] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Releasing lock "refresh_cache-0950cbd3-0708-4994-8d2d-053aecc06cdf" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.871462] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Updated the network info_cache for instance {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 633.871462] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 633.871462] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 633.871462] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 633.871462] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 633.871856] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 633.871856] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 633.871856] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63021) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 633.872034] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 633.886812] env[63021]: DEBUG nova.network.neutron [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 634.171088] env[63021]: DEBUG nova.scheduler.client.report [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 634.252839] env[63021]: DEBUG nova.compute.manager [req-668a5ae2-176e-4eff-9a0e-634a75b3b9d2 req-fd5d7483-bff7-4284-877e-cf4c6226c393 service nova] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Received event network-changed-af20fd90-4fdd-4898-9acf-ee54c4d9f27b {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 634.253117] env[63021]: DEBUG nova.compute.manager [req-668a5ae2-176e-4eff-9a0e-634a75b3b9d2 req-fd5d7483-bff7-4284-877e-cf4c6226c393 service nova] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Refreshing instance network info cache due to event network-changed-af20fd90-4fdd-4898-9acf-ee54c4d9f27b. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 634.253376] env[63021]: DEBUG oslo_concurrency.lockutils [req-668a5ae2-176e-4eff-9a0e-634a75b3b9d2 req-fd5d7483-bff7-4284-877e-cf4c6226c393 service nova] Acquiring lock "refresh_cache-ab11dead-3fe6-46d3-92e1-cdca84ebe508" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.253555] env[63021]: DEBUG oslo_concurrency.lockutils [req-668a5ae2-176e-4eff-9a0e-634a75b3b9d2 req-fd5d7483-bff7-4284-877e-cf4c6226c393 service nova] Acquired lock "refresh_cache-ab11dead-3fe6-46d3-92e1-cdca84ebe508" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.253719] env[63021]: DEBUG nova.network.neutron [req-668a5ae2-176e-4eff-9a0e-634a75b3b9d2 req-fd5d7483-bff7-4284-877e-cf4c6226c393 service nova] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Refreshing network info cache for port af20fd90-4fdd-4898-9acf-ee54c4d9f27b {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 634.286315] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Acquiring lock "3660a926-7460-41f2-9ee4-2a5072701aea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.286674] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Lock "3660a926-7460-41f2-9ee4-2a5072701aea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.325367] env[63021]: ERROR nova.compute.manager [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port af20fd90-4fdd-4898-9acf-ee54c4d9f27b, please check neutron logs for more information. [ 634.325367] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 634.325367] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.325367] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 634.325367] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 634.325367] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 634.325367] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 634.325367] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 634.325367] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.325367] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 634.325367] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.325367] env[63021]: ERROR nova.compute.manager raise self.value [ 634.325367] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 634.325367] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 634.325367] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.325367] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 634.325819] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.325819] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 634.325819] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port af20fd90-4fdd-4898-9acf-ee54c4d9f27b, please check neutron logs for more information. [ 634.325819] env[63021]: ERROR nova.compute.manager [ 634.325819] env[63021]: Traceback (most recent call last): [ 634.325819] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 634.325819] env[63021]: listener.cb(fileno) [ 634.325819] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 634.325819] env[63021]: result = function(*args, **kwargs) [ 634.325819] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 634.325819] env[63021]: return func(*args, **kwargs) [ 634.325819] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 634.325819] env[63021]: raise e [ 634.325819] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.325819] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 634.325819] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 634.325819] env[63021]: created_port_ids = self._update_ports_for_instance( [ 634.325819] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 634.325819] env[63021]: with excutils.save_and_reraise_exception(): [ 634.325819] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.325819] env[63021]: self.force_reraise() [ 634.325819] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.325819] env[63021]: raise self.value [ 634.325819] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 634.325819] env[63021]: updated_port = self._update_port( [ 634.325819] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.325819] env[63021]: _ensure_no_port_binding_failure(port) [ 634.325819] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.325819] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 634.326640] env[63021]: nova.exception.PortBindingFailed: Binding failed for port af20fd90-4fdd-4898-9acf-ee54c4d9f27b, please check neutron logs for more information. [ 634.326640] env[63021]: Removing descriptor: 17 [ 634.326640] env[63021]: ERROR nova.compute.manager [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port af20fd90-4fdd-4898-9acf-ee54c4d9f27b, please check neutron logs for more information. [ 634.326640] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Traceback (most recent call last): [ 634.326640] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 634.326640] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] yield resources [ 634.326640] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 634.326640] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] self.driver.spawn(context, instance, image_meta, [ 634.326640] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 634.326640] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] self._vmops.spawn(context, instance, image_meta, injected_files, [ 634.326640] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 634.326640] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] vm_ref = self.build_virtual_machine(instance, [ 634.327063] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 634.327063] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] vif_infos = vmwarevif.get_vif_info(self._session, [ 634.327063] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 634.327063] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] for vif in network_info: [ 634.327063] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 634.327063] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] return self._sync_wrapper(fn, *args, **kwargs) [ 634.327063] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 634.327063] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] self.wait() [ 634.327063] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 634.327063] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] self[:] = self._gt.wait() [ 634.327063] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 634.327063] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] return self._exit_event.wait() [ 634.327063] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 634.327555] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] result = hub.switch() [ 634.327555] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 634.327555] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] return self.greenlet.switch() [ 634.327555] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 634.327555] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] result = function(*args, **kwargs) [ 634.327555] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 634.327555] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] return func(*args, **kwargs) [ 634.327555] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 634.327555] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] raise e [ 634.327555] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.327555] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] nwinfo = self.network_api.allocate_for_instance( [ 634.327555] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 634.327555] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] created_port_ids = self._update_ports_for_instance( [ 634.328026] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 634.328026] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] with excutils.save_and_reraise_exception(): [ 634.328026] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.328026] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] self.force_reraise() [ 634.328026] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.328026] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] raise self.value [ 634.328026] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 634.328026] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] updated_port = self._update_port( [ 634.328026] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.328026] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] _ensure_no_port_binding_failure(port) [ 634.328026] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.328026] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] raise exception.PortBindingFailed(port_id=port['id']) [ 634.328477] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] nova.exception.PortBindingFailed: Binding failed for port af20fd90-4fdd-4898-9acf-ee54c4d9f27b, please check neutron logs for more information. [ 634.328477] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] [ 634.328477] env[63021]: INFO nova.compute.manager [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Terminating instance [ 634.330047] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Acquiring lock "refresh_cache-ab11dead-3fe6-46d3-92e1-cdca84ebe508" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.377314] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.387479] env[63021]: DEBUG nova.network.neutron [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.676703] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.992s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.677376] env[63021]: ERROR nova.compute.manager [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b5fe80a8-6652-498a-898b-2ed43886c4fd, please check neutron logs for more information. [ 634.677376] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Traceback (most recent call last): [ 634.677376] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 634.677376] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] self.driver.spawn(context, instance, image_meta, [ 634.677376] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 634.677376] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 634.677376] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 634.677376] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] vm_ref = self.build_virtual_machine(instance, [ 634.677376] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 634.677376] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] vif_infos = vmwarevif.get_vif_info(self._session, [ 634.677376] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 634.677680] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] for vif in network_info: [ 634.677680] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 634.677680] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] return self._sync_wrapper(fn, *args, **kwargs) [ 634.677680] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 634.677680] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] self.wait() [ 634.677680] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 634.677680] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] self[:] = self._gt.wait() [ 634.677680] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 634.677680] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] return self._exit_event.wait() [ 634.677680] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 634.677680] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] current.throw(*self._exc) [ 634.677680] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 634.677680] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] result = function(*args, **kwargs) [ 634.677962] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 634.677962] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] return func(*args, **kwargs) [ 634.677962] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 634.677962] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] raise e [ 634.677962] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.677962] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] nwinfo = self.network_api.allocate_for_instance( [ 634.677962] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 634.677962] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] created_port_ids = self._update_ports_for_instance( [ 634.677962] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 634.677962] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] with excutils.save_and_reraise_exception(): [ 634.677962] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.677962] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] self.force_reraise() [ 634.677962] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.678251] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] raise self.value [ 634.678251] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 634.678251] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] updated_port = self._update_port( [ 634.678251] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.678251] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] _ensure_no_port_binding_failure(port) [ 634.678251] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.678251] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] raise exception.PortBindingFailed(port_id=port['id']) [ 634.678251] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] nova.exception.PortBindingFailed: Binding failed for port b5fe80a8-6652-498a-898b-2ed43886c4fd, please check neutron logs for more information. [ 634.678251] env[63021]: ERROR nova.compute.manager [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] [ 634.678251] env[63021]: DEBUG nova.compute.utils [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Binding failed for port b5fe80a8-6652-498a-898b-2ed43886c4fd, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 634.680147] env[63021]: DEBUG nova.compute.manager [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Build of instance 17ff6693-c323-478b-8b61-776c3b41f77e was re-scheduled: Binding failed for port b5fe80a8-6652-498a-898b-2ed43886c4fd, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 634.680996] env[63021]: DEBUG nova.compute.manager [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 634.681562] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "refresh_cache-17ff6693-c323-478b-8b61-776c3b41f77e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.681825] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquired lock "refresh_cache-17ff6693-c323-478b-8b61-776c3b41f77e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.682109] env[63021]: DEBUG nova.network.neutron [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 634.683376] env[63021]: DEBUG oslo_concurrency.lockutils [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.490s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.785995] env[63021]: DEBUG nova.network.neutron [req-668a5ae2-176e-4eff-9a0e-634a75b3b9d2 req-fd5d7483-bff7-4284-877e-cf4c6226c393 service nova] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 634.890461] env[63021]: INFO nova.compute.manager [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] [instance: e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7] Took 1.02 seconds to deallocate network for instance. [ 634.946265] env[63021]: DEBUG nova.network.neutron [req-668a5ae2-176e-4eff-9a0e-634a75b3b9d2 req-fd5d7483-bff7-4284-877e-cf4c6226c393 service nova] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.209393] env[63021]: DEBUG nova.network.neutron [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 635.347432] env[63021]: DEBUG nova.network.neutron [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.449826] env[63021]: DEBUG oslo_concurrency.lockutils [req-668a5ae2-176e-4eff-9a0e-634a75b3b9d2 req-fd5d7483-bff7-4284-877e-cf4c6226c393 service nova] Releasing lock "refresh_cache-ab11dead-3fe6-46d3-92e1-cdca84ebe508" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 635.450227] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Acquired lock "refresh_cache-ab11dead-3fe6-46d3-92e1-cdca84ebe508" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.450408] env[63021]: DEBUG nova.network.neutron [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 635.647418] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ced3ab23-723b-4b0d-9a5d-dffe41720e63 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.659031] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70d49e54-61f5-4282-8a8c-3184f094a3aa {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.698459] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d9c5628-51d0-474f-94f2-f62e603965fb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.710179] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b0f5b9b-b47a-4fa7-8652-1d82c6e79d06 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.722930] env[63021]: DEBUG nova.compute.provider_tree [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 635.853530] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Releasing lock "refresh_cache-17ff6693-c323-478b-8b61-776c3b41f77e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 635.853829] env[63021]: DEBUG nova.compute.manager [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 635.854271] env[63021]: DEBUG nova.compute.manager [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 635.854455] env[63021]: DEBUG nova.network.neutron [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 635.886414] env[63021]: DEBUG nova.network.neutron [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 635.933113] env[63021]: INFO nova.scheduler.client.report [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Deleted allocations for instance e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7 [ 635.985028] env[63021]: DEBUG nova.network.neutron [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 636.191164] env[63021]: DEBUG nova.network.neutron [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.226254] env[63021]: DEBUG nova.scheduler.client.report [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 636.317753] env[63021]: DEBUG nova.compute.manager [req-eb5723bc-cc99-4e16-97c4-9e5575afc4e3 req-ca449713-de8e-479e-95fb-56dec9bad898 service nova] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Received event network-vif-deleted-af20fd90-4fdd-4898-9acf-ee54c4d9f27b {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 636.392454] env[63021]: DEBUG nova.network.neutron [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.443197] env[63021]: DEBUG oslo_concurrency.lockutils [None req-51cb2b6b-1adf-43c3-b488-473ce019bb99 tempest-FloatingIPsAssociationNegativeTestJSON-996113527 tempest-FloatingIPsAssociationNegativeTestJSON-996113527-project-member] Lock "e4615bf3-88d4-44b2-bd31-ef9d85ba5ae7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.601s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 636.695147] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Releasing lock "refresh_cache-ab11dead-3fe6-46d3-92e1-cdca84ebe508" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.695147] env[63021]: DEBUG nova.compute.manager [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 636.695147] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 636.697194] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-95cbfb6b-ca07-4adf-8645-0f8390a3cfbd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.707489] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01cbe1a0-c1df-4919-bbde-436a3985a89b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.731535] env[63021]: DEBUG oslo_concurrency.lockutils [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.048s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 636.732145] env[63021]: ERROR nova.compute.manager [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8b1b2b2c-fec7-4c23-8349-bc89d9cc5b36, please check neutron logs for more information. [ 636.732145] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Traceback (most recent call last): [ 636.732145] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 636.732145] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] self.driver.spawn(context, instance, image_meta, [ 636.732145] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 636.732145] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 636.732145] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 636.732145] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] vm_ref = self.build_virtual_machine(instance, [ 636.732145] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 636.732145] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] vif_infos = vmwarevif.get_vif_info(self._session, [ 636.732145] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 636.732466] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] for vif in network_info: [ 636.732466] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 636.732466] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] return self._sync_wrapper(fn, *args, **kwargs) [ 636.732466] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 636.732466] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] self.wait() [ 636.732466] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 636.732466] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] self[:] = self._gt.wait() [ 636.732466] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 636.732466] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] return self._exit_event.wait() [ 636.732466] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 636.732466] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] current.throw(*self._exc) [ 636.732466] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 636.732466] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] result = function(*args, **kwargs) [ 636.732743] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 636.732743] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] return func(*args, **kwargs) [ 636.732743] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 636.732743] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] raise e [ 636.732743] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 636.732743] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] nwinfo = self.network_api.allocate_for_instance( [ 636.732743] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 636.732743] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] created_port_ids = self._update_ports_for_instance( [ 636.732743] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 636.732743] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] with excutils.save_and_reraise_exception(): [ 636.732743] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.732743] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] self.force_reraise() [ 636.732743] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.733039] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] raise self.value [ 636.733039] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 636.733039] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] updated_port = self._update_port( [ 636.733039] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.733039] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] _ensure_no_port_binding_failure(port) [ 636.733039] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.733039] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] raise exception.PortBindingFailed(port_id=port['id']) [ 636.733039] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] nova.exception.PortBindingFailed: Binding failed for port 8b1b2b2c-fec7-4c23-8349-bc89d9cc5b36, please check neutron logs for more information. [ 636.733039] env[63021]: ERROR nova.compute.manager [instance: efd0e375-3134-4984-8f82-48b53c65e26c] [ 636.734037] env[63021]: DEBUG nova.compute.utils [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Binding failed for port 8b1b2b2c-fec7-4c23-8349-bc89d9cc5b36, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 636.740155] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.933s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.744402] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ab11dead-3fe6-46d3-92e1-cdca84ebe508 could not be found. [ 636.744402] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 636.744402] env[63021]: INFO nova.compute.manager [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Took 0.05 seconds to destroy the instance on the hypervisor. [ 636.744402] env[63021]: DEBUG oslo.service.loopingcall [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 636.745018] env[63021]: DEBUG nova.compute.manager [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Build of instance efd0e375-3134-4984-8f82-48b53c65e26c was re-scheduled: Binding failed for port 8b1b2b2c-fec7-4c23-8349-bc89d9cc5b36, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 636.745952] env[63021]: DEBUG nova.compute.manager [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 636.746343] env[63021]: DEBUG oslo_concurrency.lockutils [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Acquiring lock "refresh_cache-efd0e375-3134-4984-8f82-48b53c65e26c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.746529] env[63021]: DEBUG oslo_concurrency.lockutils [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Acquired lock "refresh_cache-efd0e375-3134-4984-8f82-48b53c65e26c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.746778] env[63021]: DEBUG nova.network.neutron [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 636.748782] env[63021]: DEBUG nova.compute.manager [-] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 636.748782] env[63021]: DEBUG nova.network.neutron [-] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 636.791559] env[63021]: DEBUG nova.network.neutron [-] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 636.895032] env[63021]: INFO nova.compute.manager [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 17ff6693-c323-478b-8b61-776c3b41f77e] Took 1.04 seconds to deallocate network for instance. [ 636.947504] env[63021]: DEBUG nova.compute.manager [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 637.278335] env[63021]: DEBUG nova.network.neutron [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.296037] env[63021]: DEBUG nova.network.neutron [-] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.367533] env[63021]: DEBUG nova.network.neutron [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.479025] env[63021]: DEBUG oslo_concurrency.lockutils [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.700649] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7942ae17-e692-4a48-b12d-f3564cb3b11f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.706845] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b32fa282-a27f-495b-ac03-c9b5f6e33e13 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.740988] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce12a119-cc79-44e6-85bc-944b3b626b45 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.750808] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6280fb9e-7efc-43bd-91cc-6dec55f14fc4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.766100] env[63021]: DEBUG nova.compute.provider_tree [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 637.800540] env[63021]: INFO nova.compute.manager [-] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Took 1.05 seconds to deallocate network for instance. [ 637.805266] env[63021]: DEBUG nova.compute.claims [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 637.805812] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.873555] env[63021]: DEBUG oslo_concurrency.lockutils [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Releasing lock "refresh_cache-efd0e375-3134-4984-8f82-48b53c65e26c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.873555] env[63021]: DEBUG nova.compute.manager [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 637.873723] env[63021]: DEBUG nova.compute.manager [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 637.873761] env[63021]: DEBUG nova.network.neutron [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 637.900782] env[63021]: DEBUG nova.network.neutron [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.948375] env[63021]: INFO nova.scheduler.client.report [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Deleted allocations for instance 17ff6693-c323-478b-8b61-776c3b41f77e [ 638.270037] env[63021]: DEBUG nova.scheduler.client.report [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 638.406769] env[63021]: DEBUG nova.network.neutron [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.456677] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9a3dc9ae-84b0-44e8-9f5b-a9b10e6ab4b3 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "17ff6693-c323-478b-8b61-776c3b41f77e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 75.507s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 638.775353] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.035s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 638.776411] env[63021]: ERROR nova.compute.manager [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c25b24ca-5c74-4626-9705-e6c28e49a8c5, please check neutron logs for more information. [ 638.776411] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Traceback (most recent call last): [ 638.776411] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 638.776411] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] self.driver.spawn(context, instance, image_meta, [ 638.776411] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 638.776411] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 638.776411] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 638.776411] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] vm_ref = self.build_virtual_machine(instance, [ 638.776411] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 638.776411] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] vif_infos = vmwarevif.get_vif_info(self._session, [ 638.776411] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 638.776761] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] for vif in network_info: [ 638.776761] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 638.776761] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] return self._sync_wrapper(fn, *args, **kwargs) [ 638.776761] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 638.776761] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] self.wait() [ 638.776761] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 638.776761] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] self[:] = self._gt.wait() [ 638.776761] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 638.776761] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] return self._exit_event.wait() [ 638.776761] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 638.776761] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] current.throw(*self._exc) [ 638.776761] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.776761] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] result = function(*args, **kwargs) [ 638.777107] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 638.777107] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] return func(*args, **kwargs) [ 638.777107] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 638.777107] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] raise e [ 638.777107] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.777107] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] nwinfo = self.network_api.allocate_for_instance( [ 638.777107] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 638.777107] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] created_port_ids = self._update_ports_for_instance( [ 638.777107] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 638.777107] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] with excutils.save_and_reraise_exception(): [ 638.777107] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.777107] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] self.force_reraise() [ 638.777107] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.777451] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] raise self.value [ 638.777451] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 638.777451] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] updated_port = self._update_port( [ 638.777451] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.777451] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] _ensure_no_port_binding_failure(port) [ 638.777451] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.777451] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] raise exception.PortBindingFailed(port_id=port['id']) [ 638.777451] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] nova.exception.PortBindingFailed: Binding failed for port c25b24ca-5c74-4626-9705-e6c28e49a8c5, please check neutron logs for more information. [ 638.777451] env[63021]: ERROR nova.compute.manager [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] [ 638.777451] env[63021]: DEBUG nova.compute.utils [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Binding failed for port c25b24ca-5c74-4626-9705-e6c28e49a8c5, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 638.778620] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.563s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.781688] env[63021]: DEBUG nova.compute.manager [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Build of instance 92280cdf-bad2-4893-9a37-706cb1bcbac4 was re-scheduled: Binding failed for port c25b24ca-5c74-4626-9705-e6c28e49a8c5, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 638.782233] env[63021]: DEBUG nova.compute.manager [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 638.782473] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Acquiring lock "refresh_cache-92280cdf-bad2-4893-9a37-706cb1bcbac4" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.782619] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Acquired lock "refresh_cache-92280cdf-bad2-4893-9a37-706cb1bcbac4" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.782773] env[63021]: DEBUG nova.network.neutron [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 638.909245] env[63021]: INFO nova.compute.manager [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] [instance: efd0e375-3134-4984-8f82-48b53c65e26c] Took 1.04 seconds to deallocate network for instance. [ 638.963023] env[63021]: DEBUG nova.compute.manager [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 639.337091] env[63021]: DEBUG nova.network.neutron [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.504830] env[63021]: DEBUG oslo_concurrency.lockutils [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.554938] env[63021]: DEBUG nova.network.neutron [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.764514] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da8373bb-498f-4e06-924a-c6e505bcb4cb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.773309] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00684f22-7082-4dd7-bd49-289fc7787f91 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.815168] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e43dc748-d46b-4ee5-8100-ce81095dd4aa {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.824589] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d000ddf1-f244-40ec-b0b6-fb08983f8288 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.848511] env[63021]: DEBUG nova.compute.provider_tree [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 639.929806] env[63021]: DEBUG oslo_concurrency.lockutils [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Acquiring lock "bc48ac80-36f6-4c97-9f95-29aed2efeae6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.929806] env[63021]: DEBUG oslo_concurrency.lockutils [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Lock "bc48ac80-36f6-4c97-9f95-29aed2efeae6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.973253] env[63021]: INFO nova.scheduler.client.report [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Deleted allocations for instance efd0e375-3134-4984-8f82-48b53c65e26c [ 640.057774] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Releasing lock "refresh_cache-92280cdf-bad2-4893-9a37-706cb1bcbac4" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.061016] env[63021]: DEBUG nova.compute.manager [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 640.061016] env[63021]: DEBUG nova.compute.manager [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 640.061016] env[63021]: DEBUG nova.network.neutron [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 640.085101] env[63021]: DEBUG nova.network.neutron [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 640.355514] env[63021]: DEBUG nova.scheduler.client.report [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 640.484017] env[63021]: DEBUG oslo_concurrency.lockutils [None req-57d4ca9c-080f-4448-898f-f9089b0f12f6 tempest-DeleteServersAdminTestJSON-1829816874 tempest-DeleteServersAdminTestJSON-1829816874-project-member] Lock "efd0e375-3134-4984-8f82-48b53c65e26c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.225s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.587040] env[63021]: DEBUG nova.network.neutron [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.774020] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "907540da-a701-477a-9b5c-9942b5d2d987" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.774020] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "907540da-a701-477a-9b5c-9942b5d2d987" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.865734] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.086s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.865734] env[63021]: ERROR nova.compute.manager [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 758c2831-cb51-46c3-a612-cbbd0215f383, please check neutron logs for more information. [ 640.865734] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Traceback (most recent call last): [ 640.865734] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 640.865734] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] self.driver.spawn(context, instance, image_meta, [ 640.865734] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 640.865734] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 640.865734] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 640.865734] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] vm_ref = self.build_virtual_machine(instance, [ 640.866067] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 640.866067] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] vif_infos = vmwarevif.get_vif_info(self._session, [ 640.866067] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 640.866067] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] for vif in network_info: [ 640.866067] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 640.866067] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] return self._sync_wrapper(fn, *args, **kwargs) [ 640.866067] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 640.866067] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] self.wait() [ 640.866067] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 640.866067] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] self[:] = self._gt.wait() [ 640.866067] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 640.866067] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] return self._exit_event.wait() [ 640.866067] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 640.866448] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] result = hub.switch() [ 640.866448] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 640.866448] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] return self.greenlet.switch() [ 640.866448] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 640.866448] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] result = function(*args, **kwargs) [ 640.866448] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 640.866448] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] return func(*args, **kwargs) [ 640.866448] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 640.866448] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] raise e [ 640.866448] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 640.866448] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] nwinfo = self.network_api.allocate_for_instance( [ 640.866448] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 640.866448] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] created_port_ids = self._update_ports_for_instance( [ 640.866734] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 640.866734] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] with excutils.save_and_reraise_exception(): [ 640.866734] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.866734] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] self.force_reraise() [ 640.866734] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.866734] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] raise self.value [ 640.866734] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 640.866734] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] updated_port = self._update_port( [ 640.866734] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.866734] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] _ensure_no_port_binding_failure(port) [ 640.866734] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.866734] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] raise exception.PortBindingFailed(port_id=port['id']) [ 640.867047] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] nova.exception.PortBindingFailed: Binding failed for port 758c2831-cb51-46c3-a612-cbbd0215f383, please check neutron logs for more information. [ 640.867047] env[63021]: ERROR nova.compute.manager [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] [ 640.867047] env[63021]: DEBUG nova.compute.utils [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Binding failed for port 758c2831-cb51-46c3-a612-cbbd0215f383, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 640.867761] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.159s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.867998] env[63021]: DEBUG nova.objects.instance [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Lazy-loading 'resources' on Instance uuid 0950cbd3-0708-4994-8d2d-053aecc06cdf {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 640.869500] env[63021]: DEBUG nova.compute.manager [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Build of instance cee88fa5-c98e-405f-b41c-62d6661b983c was re-scheduled: Binding failed for port 758c2831-cb51-46c3-a612-cbbd0215f383, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 640.869933] env[63021]: DEBUG nova.compute.manager [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 640.870165] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Acquiring lock "refresh_cache-cee88fa5-c98e-405f-b41c-62d6661b983c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 640.870311] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Acquired lock "refresh_cache-cee88fa5-c98e-405f-b41c-62d6661b983c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.870467] env[63021]: DEBUG nova.network.neutron [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 640.884728] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Acquiring lock "06ef77f1-a196-499c-b639-8427b021a8c0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.884728] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lock "06ef77f1-a196-499c-b639-8427b021a8c0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.984790] env[63021]: DEBUG nova.compute.manager [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 641.089447] env[63021]: INFO nova.compute.manager [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] [instance: 92280cdf-bad2-4893-9a37-706cb1bcbac4] Took 1.03 seconds to deallocate network for instance. [ 641.418729] env[63021]: DEBUG nova.network.neutron [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 641.514836] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.645074] env[63021]: DEBUG nova.network.neutron [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.773944] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5847e71b-7b72-41d2-95c1-b2ade57d071d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.782459] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ec0f5df-b261-4a0b-8bf1-de0bfba5189b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.812603] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7eee0c8-b103-4ac2-9588-8427ce72b5f0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.820300] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e82500f5-7a67-4798-a280-212282194d25 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.837123] env[63021]: DEBUG nova.compute.provider_tree [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 642.130428] env[63021]: INFO nova.scheduler.client.report [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Deleted allocations for instance 92280cdf-bad2-4893-9a37-706cb1bcbac4 [ 642.153380] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Releasing lock "refresh_cache-cee88fa5-c98e-405f-b41c-62d6661b983c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.153609] env[63021]: DEBUG nova.compute.manager [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 642.153774] env[63021]: DEBUG nova.compute.manager [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 642.153930] env[63021]: DEBUG nova.network.neutron [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 642.172323] env[63021]: DEBUG nova.network.neutron [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.340755] env[63021]: DEBUG nova.scheduler.client.report [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 642.640226] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b8523a69-dad1-4b18-b0c5-da82cdad22f0 tempest-ImagesOneServerNegativeTestJSON-1414078702 tempest-ImagesOneServerNegativeTestJSON-1414078702-project-member] Lock "92280cdf-bad2-4893-9a37-706cb1bcbac4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.287s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.676048] env[63021]: DEBUG nova.network.neutron [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.847933] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.978s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.848956] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.964s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.851607] env[63021]: INFO nova.compute.claims [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 642.881623] env[63021]: INFO nova.scheduler.client.report [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Deleted allocations for instance 0950cbd3-0708-4994-8d2d-053aecc06cdf [ 643.145423] env[63021]: DEBUG nova.compute.manager [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 643.179053] env[63021]: INFO nova.compute.manager [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: cee88fa5-c98e-405f-b41c-62d6661b983c] Took 1.02 seconds to deallocate network for instance. [ 643.392063] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2a18cd7c-d6c3-49fc-92e2-2c9593777158 tempest-ServerShowV247Test-403622186 tempest-ServerShowV247Test-403622186-project-member] Lock "0950cbd3-0708-4994-8d2d-053aecc06cdf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.360s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 643.676449] env[63021]: DEBUG oslo_concurrency.lockutils [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 644.218669] env[63021]: INFO nova.scheduler.client.report [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Deleted allocations for instance cee88fa5-c98e-405f-b41c-62d6661b983c [ 644.334048] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7285666c-2649-4cfe-a60d-bd4aec7b13a5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.343434] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4233adc-30ac-4f99-b7f9-83b591ed610b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.382248] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51ae460a-09a0-4b98-a9fa-7490285f4497 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.390840] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f65bf8e4-fea4-4861-9d2f-2bac5fcea8ab {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.405184] env[63021]: DEBUG nova.compute.provider_tree [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 644.734670] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f57edf29-d66f-46cd-b54e-e768283e2393 tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Lock "cee88fa5-c98e-405f-b41c-62d6661b983c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.960s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.914449] env[63021]: DEBUG nova.scheduler.client.report [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 645.237665] env[63021]: DEBUG nova.compute.manager [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 645.421949] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.570s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.421949] env[63021]: DEBUG nova.compute.manager [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 645.423033] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.542s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.424798] env[63021]: INFO nova.compute.claims [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 645.736255] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Acquiring lock "8ea2a693-0cd8-4f91-8bd8-b40265b9372e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.736487] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Lock "8ea2a693-0cd8-4f91-8bd8-b40265b9372e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.762356] env[63021]: DEBUG oslo_concurrency.lockutils [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.930190] env[63021]: DEBUG nova.compute.utils [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 645.931227] env[63021]: DEBUG nova.compute.manager [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 645.931462] env[63021]: DEBUG nova.network.neutron [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 646.021504] env[63021]: DEBUG nova.policy [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1e133aeda3249ec98dbdc53f869d719', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7331caed42034dfc81c79cfabbe0829e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 646.437542] env[63021]: DEBUG nova.compute.manager [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 646.519953] env[63021]: DEBUG nova.network.neutron [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Successfully created port: 97e064ac-be35-4a14-960e-34af1585617f {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 646.870867] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b04c65d1-3d43-46ca-ac3f-f5ab9ffd358b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.883137] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7e9514c-4821-4a72-9190-30da28d6b012 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.916955] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01f5d92b-5bca-4285-a83e-f2d3d40b7dc3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.924614] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cbb5d13-c9b5-4dea-bb52-7383bdc514b0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.940726] env[63021]: DEBUG nova.compute.provider_tree [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 647.441520] env[63021]: DEBUG nova.scheduler.client.report [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 647.449477] env[63021]: DEBUG nova.compute.manager [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 647.482016] env[63021]: DEBUG nova.virt.hardware [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 647.482016] env[63021]: DEBUG nova.virt.hardware [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 647.482016] env[63021]: DEBUG nova.virt.hardware [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 647.482016] env[63021]: DEBUG nova.virt.hardware [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 647.482397] env[63021]: DEBUG nova.virt.hardware [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 647.482818] env[63021]: DEBUG nova.virt.hardware [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 647.486017] env[63021]: DEBUG nova.virt.hardware [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 647.486017] env[63021]: DEBUG nova.virt.hardware [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 647.486017] env[63021]: DEBUG nova.virt.hardware [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 647.486017] env[63021]: DEBUG nova.virt.hardware [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 647.486017] env[63021]: DEBUG nova.virt.hardware [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 647.486017] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77c5268e-133b-4a57-bdc4-5020d9ddaa9c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.497349] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6273dc98-0eea-4d10-a26d-c9ed46dbe980 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.948273] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.525s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.948963] env[63021]: DEBUG nova.compute.manager [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 647.952028] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.236s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.953445] env[63021]: INFO nova.compute.claims [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 648.396669] env[63021]: DEBUG nova.compute.manager [req-0ff0e19c-6b6c-40a3-9d47-a510344458b2 req-74c2b058-ed5f-4f27-af0b-4375cf042cd1 service nova] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Received event network-changed-97e064ac-be35-4a14-960e-34af1585617f {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 648.396860] env[63021]: DEBUG nova.compute.manager [req-0ff0e19c-6b6c-40a3-9d47-a510344458b2 req-74c2b058-ed5f-4f27-af0b-4375cf042cd1 service nova] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Refreshing instance network info cache due to event network-changed-97e064ac-be35-4a14-960e-34af1585617f. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 648.397176] env[63021]: DEBUG oslo_concurrency.lockutils [req-0ff0e19c-6b6c-40a3-9d47-a510344458b2 req-74c2b058-ed5f-4f27-af0b-4375cf042cd1 service nova] Acquiring lock "refresh_cache-12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 648.397220] env[63021]: DEBUG oslo_concurrency.lockutils [req-0ff0e19c-6b6c-40a3-9d47-a510344458b2 req-74c2b058-ed5f-4f27-af0b-4375cf042cd1 service nova] Acquired lock "refresh_cache-12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.397736] env[63021]: DEBUG nova.network.neutron [req-0ff0e19c-6b6c-40a3-9d47-a510344458b2 req-74c2b058-ed5f-4f27-af0b-4375cf042cd1 service nova] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Refreshing network info cache for port 97e064ac-be35-4a14-960e-34af1585617f {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 648.464022] env[63021]: DEBUG nova.compute.utils [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 648.468083] env[63021]: DEBUG nova.compute.manager [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 648.468083] env[63021]: DEBUG nova.network.neutron [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 648.513281] env[63021]: ERROR nova.compute.manager [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 97e064ac-be35-4a14-960e-34af1585617f, please check neutron logs for more information. [ 648.513281] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 648.513281] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.513281] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 648.513281] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 648.513281] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 648.513281] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 648.513281] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 648.513281] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.513281] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 648.513281] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.513281] env[63021]: ERROR nova.compute.manager raise self.value [ 648.513281] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 648.513281] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 648.513281] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.513281] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 648.513701] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.513701] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 648.513701] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 97e064ac-be35-4a14-960e-34af1585617f, please check neutron logs for more information. [ 648.513701] env[63021]: ERROR nova.compute.manager [ 648.513701] env[63021]: Traceback (most recent call last): [ 648.513701] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 648.513701] env[63021]: listener.cb(fileno) [ 648.513701] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 648.513701] env[63021]: result = function(*args, **kwargs) [ 648.513701] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 648.513701] env[63021]: return func(*args, **kwargs) [ 648.513701] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 648.513701] env[63021]: raise e [ 648.513701] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.513701] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 648.513701] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 648.513701] env[63021]: created_port_ids = self._update_ports_for_instance( [ 648.513701] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 648.513701] env[63021]: with excutils.save_and_reraise_exception(): [ 648.513701] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.513701] env[63021]: self.force_reraise() [ 648.513701] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.513701] env[63021]: raise self.value [ 648.513701] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 648.513701] env[63021]: updated_port = self._update_port( [ 648.513701] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.513701] env[63021]: _ensure_no_port_binding_failure(port) [ 648.513701] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.513701] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 648.514386] env[63021]: nova.exception.PortBindingFailed: Binding failed for port 97e064ac-be35-4a14-960e-34af1585617f, please check neutron logs for more information. [ 648.514386] env[63021]: Removing descriptor: 17 [ 648.514386] env[63021]: ERROR nova.compute.manager [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 97e064ac-be35-4a14-960e-34af1585617f, please check neutron logs for more information. [ 648.514386] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Traceback (most recent call last): [ 648.514386] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 648.514386] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] yield resources [ 648.514386] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 648.514386] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] self.driver.spawn(context, instance, image_meta, [ 648.514386] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 648.514386] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 648.514386] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 648.514386] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] vm_ref = self.build_virtual_machine(instance, [ 648.514823] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 648.514823] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] vif_infos = vmwarevif.get_vif_info(self._session, [ 648.514823] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 648.514823] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] for vif in network_info: [ 648.514823] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 648.514823] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] return self._sync_wrapper(fn, *args, **kwargs) [ 648.514823] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 648.514823] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] self.wait() [ 648.514823] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 648.514823] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] self[:] = self._gt.wait() [ 648.514823] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 648.514823] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] return self._exit_event.wait() [ 648.514823] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 648.515120] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] result = hub.switch() [ 648.515120] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 648.515120] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] return self.greenlet.switch() [ 648.515120] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 648.515120] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] result = function(*args, **kwargs) [ 648.515120] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 648.515120] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] return func(*args, **kwargs) [ 648.515120] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 648.515120] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] raise e [ 648.515120] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.515120] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] nwinfo = self.network_api.allocate_for_instance( [ 648.515120] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 648.515120] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] created_port_ids = self._update_ports_for_instance( [ 648.515506] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 648.515506] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] with excutils.save_and_reraise_exception(): [ 648.515506] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.515506] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] self.force_reraise() [ 648.515506] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.515506] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] raise self.value [ 648.515506] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 648.515506] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] updated_port = self._update_port( [ 648.515506] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.515506] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] _ensure_no_port_binding_failure(port) [ 648.515506] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.515506] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] raise exception.PortBindingFailed(port_id=port['id']) [ 648.515844] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] nova.exception.PortBindingFailed: Binding failed for port 97e064ac-be35-4a14-960e-34af1585617f, please check neutron logs for more information. [ 648.515844] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] [ 648.515844] env[63021]: INFO nova.compute.manager [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Terminating instance [ 648.518105] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "refresh_cache-12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 648.556394] env[63021]: DEBUG nova.policy [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1174037b1714a6dbd14de4aed339444', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '758a30d0205d4534b767f42c22591ef5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 648.955122] env[63021]: DEBUG nova.network.neutron [req-0ff0e19c-6b6c-40a3-9d47-a510344458b2 req-74c2b058-ed5f-4f27-af0b-4375cf042cd1 service nova] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 648.960349] env[63021]: DEBUG nova.network.neutron [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Successfully created port: 0b0dcc76-2a8c-4f06-b32e-22e39f072235 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 648.969837] env[63021]: DEBUG nova.compute.manager [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 649.192989] env[63021]: DEBUG nova.network.neutron [req-0ff0e19c-6b6c-40a3-9d47-a510344458b2 req-74c2b058-ed5f-4f27-af0b-4375cf042cd1 service nova] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.406577] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a592918c-e972-4ed9-b1df-e95623b91ea8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.418821] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f01a72db-ce6b-474a-b698-ca2432604a1f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.460272] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-861250a1-9598-4243-81a6-09b94652204b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.470674] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecf6d31e-c37e-4fbf-ad0f-b57849de37ca {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.498184] env[63021]: DEBUG nova.compute.provider_tree [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 649.695960] env[63021]: DEBUG oslo_concurrency.lockutils [req-0ff0e19c-6b6c-40a3-9d47-a510344458b2 req-74c2b058-ed5f-4f27-af0b-4375cf042cd1 service nova] Releasing lock "refresh_cache-12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 649.696416] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquired lock "refresh_cache-12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 649.696605] env[63021]: DEBUG nova.network.neutron [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 649.927014] env[63021]: ERROR nova.compute.manager [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0b0dcc76-2a8c-4f06-b32e-22e39f072235, please check neutron logs for more information. [ 649.927014] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 649.927014] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 649.927014] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 649.927014] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 649.927014] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 649.927014] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 649.927014] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 649.927014] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 649.927014] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 649.927014] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 649.927014] env[63021]: ERROR nova.compute.manager raise self.value [ 649.927014] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 649.927014] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 649.927014] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 649.927014] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 649.927479] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 649.927479] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 649.927479] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0b0dcc76-2a8c-4f06-b32e-22e39f072235, please check neutron logs for more information. [ 649.927479] env[63021]: ERROR nova.compute.manager [ 649.927479] env[63021]: Traceback (most recent call last): [ 649.927479] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 649.927479] env[63021]: listener.cb(fileno) [ 649.927479] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 649.927479] env[63021]: result = function(*args, **kwargs) [ 649.927479] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 649.927479] env[63021]: return func(*args, **kwargs) [ 649.927479] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 649.927479] env[63021]: raise e [ 649.927479] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 649.927479] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 649.927479] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 649.927479] env[63021]: created_port_ids = self._update_ports_for_instance( [ 649.927479] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 649.927479] env[63021]: with excutils.save_and_reraise_exception(): [ 649.927479] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 649.927479] env[63021]: self.force_reraise() [ 649.927479] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 649.927479] env[63021]: raise self.value [ 649.927479] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 649.927479] env[63021]: updated_port = self._update_port( [ 649.927479] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 649.927479] env[63021]: _ensure_no_port_binding_failure(port) [ 649.927479] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 649.927479] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 649.928247] env[63021]: nova.exception.PortBindingFailed: Binding failed for port 0b0dcc76-2a8c-4f06-b32e-22e39f072235, please check neutron logs for more information. [ 649.928247] env[63021]: Removing descriptor: 18 [ 649.985390] env[63021]: DEBUG nova.compute.manager [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 650.000249] env[63021]: DEBUG nova.scheduler.client.report [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 650.022483] env[63021]: DEBUG nova.virt.hardware [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 650.022668] env[63021]: DEBUG nova.virt.hardware [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 650.022817] env[63021]: DEBUG nova.virt.hardware [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 650.022989] env[63021]: DEBUG nova.virt.hardware [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 650.023152] env[63021]: DEBUG nova.virt.hardware [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 650.023296] env[63021]: DEBUG nova.virt.hardware [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 650.023501] env[63021]: DEBUG nova.virt.hardware [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 650.023648] env[63021]: DEBUG nova.virt.hardware [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 650.023828] env[63021]: DEBUG nova.virt.hardware [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 650.023997] env[63021]: DEBUG nova.virt.hardware [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 650.024173] env[63021]: DEBUG nova.virt.hardware [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 650.025269] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3041da48-dabe-4ad0-9df9-e5db9686eb88 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.035558] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63e0f1e9-dbb0-4545-abed-bfc4fd55166c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.053547] env[63021]: ERROR nova.compute.manager [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0b0dcc76-2a8c-4f06-b32e-22e39f072235, please check neutron logs for more information. [ 650.053547] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Traceback (most recent call last): [ 650.053547] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 650.053547] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] yield resources [ 650.053547] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 650.053547] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] self.driver.spawn(context, instance, image_meta, [ 650.053547] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 650.053547] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] self._vmops.spawn(context, instance, image_meta, injected_files, [ 650.053547] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 650.053547] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] vm_ref = self.build_virtual_machine(instance, [ 650.053547] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 650.054074] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] vif_infos = vmwarevif.get_vif_info(self._session, [ 650.054074] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 650.054074] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] for vif in network_info: [ 650.054074] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 650.054074] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] return self._sync_wrapper(fn, *args, **kwargs) [ 650.054074] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 650.054074] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] self.wait() [ 650.054074] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 650.054074] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] self[:] = self._gt.wait() [ 650.054074] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 650.054074] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] return self._exit_event.wait() [ 650.054074] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 650.054074] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] current.throw(*self._exc) [ 650.054950] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 650.054950] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] result = function(*args, **kwargs) [ 650.054950] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 650.054950] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] return func(*args, **kwargs) [ 650.054950] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 650.054950] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] raise e [ 650.054950] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 650.054950] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] nwinfo = self.network_api.allocate_for_instance( [ 650.054950] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 650.054950] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] created_port_ids = self._update_ports_for_instance( [ 650.054950] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 650.054950] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] with excutils.save_and_reraise_exception(): [ 650.054950] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 650.055376] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] self.force_reraise() [ 650.055376] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 650.055376] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] raise self.value [ 650.055376] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 650.055376] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] updated_port = self._update_port( [ 650.055376] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 650.055376] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] _ensure_no_port_binding_failure(port) [ 650.055376] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 650.055376] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] raise exception.PortBindingFailed(port_id=port['id']) [ 650.055376] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] nova.exception.PortBindingFailed: Binding failed for port 0b0dcc76-2a8c-4f06-b32e-22e39f072235, please check neutron logs for more information. [ 650.055376] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] [ 650.055376] env[63021]: INFO nova.compute.manager [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Terminating instance [ 650.055733] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Acquiring lock "refresh_cache-7da1d0f4-5dbe-4243-ac87-e476d5870882" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.055733] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Acquired lock "refresh_cache-7da1d0f4-5dbe-4243-ac87-e476d5870882" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.055733] env[63021]: DEBUG nova.network.neutron [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 650.216298] env[63021]: DEBUG nova.network.neutron [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 650.332420] env[63021]: DEBUG nova.network.neutron [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.436330] env[63021]: DEBUG nova.compute.manager [req-c8118252-8a68-4436-9716-dc10fefc992f req-95615165-a312-411b-ab80-32362db4d702 service nova] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Received event network-changed-0b0dcc76-2a8c-4f06-b32e-22e39f072235 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 650.436330] env[63021]: DEBUG nova.compute.manager [req-c8118252-8a68-4436-9716-dc10fefc992f req-95615165-a312-411b-ab80-32362db4d702 service nova] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Refreshing instance network info cache due to event network-changed-0b0dcc76-2a8c-4f06-b32e-22e39f072235. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 650.436330] env[63021]: DEBUG oslo_concurrency.lockutils [req-c8118252-8a68-4436-9716-dc10fefc992f req-95615165-a312-411b-ab80-32362db4d702 service nova] Acquiring lock "refresh_cache-7da1d0f4-5dbe-4243-ac87-e476d5870882" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.505971] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.554s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.506508] env[63021]: DEBUG nova.compute.manager [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 650.508956] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.487s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 650.510954] env[63021]: INFO nova.compute.claims [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 650.574732] env[63021]: DEBUG nova.network.neutron [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 650.656686] env[63021]: DEBUG nova.network.neutron [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.834979] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Releasing lock "refresh_cache-12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 650.835624] env[63021]: DEBUG nova.compute.manager [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 650.835840] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 650.836153] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9d7335d6-0b54-43f1-9ac2-bf73da282386 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.847043] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ad8c5bc-28a8-4dd1-9ac4-045e86de734a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.863161] env[63021]: DEBUG nova.compute.manager [req-8059b822-a86f-49c6-96b3-a31bb979ba76 req-ad2e61ef-35a7-4f91-b58a-35a782ce6dbf service nova] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Received event network-vif-deleted-97e064ac-be35-4a14-960e-34af1585617f {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 650.869319] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c could not be found. [ 650.869627] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 650.869850] env[63021]: INFO nova.compute.manager [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Took 0.03 seconds to destroy the instance on the hypervisor. [ 650.870146] env[63021]: DEBUG oslo.service.loopingcall [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 650.870486] env[63021]: DEBUG nova.compute.manager [-] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 650.870624] env[63021]: DEBUG nova.network.neutron [-] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 650.888107] env[63021]: DEBUG nova.network.neutron [-] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 651.015483] env[63021]: DEBUG nova.compute.utils [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 651.020458] env[63021]: DEBUG nova.compute.manager [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 651.020865] env[63021]: DEBUG nova.network.neutron [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 651.079539] env[63021]: DEBUG nova.policy [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1174037b1714a6dbd14de4aed339444', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '758a30d0205d4534b767f42c22591ef5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 651.159701] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Releasing lock "refresh_cache-7da1d0f4-5dbe-4243-ac87-e476d5870882" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.160510] env[63021]: DEBUG nova.compute.manager [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 651.160510] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 651.160824] env[63021]: DEBUG oslo_concurrency.lockutils [req-c8118252-8a68-4436-9716-dc10fefc992f req-95615165-a312-411b-ab80-32362db4d702 service nova] Acquired lock "refresh_cache-7da1d0f4-5dbe-4243-ac87-e476d5870882" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.161388] env[63021]: DEBUG nova.network.neutron [req-c8118252-8a68-4436-9716-dc10fefc992f req-95615165-a312-411b-ab80-32362db4d702 service nova] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Refreshing network info cache for port 0b0dcc76-2a8c-4f06-b32e-22e39f072235 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 651.162549] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-71c7a02e-2c82-41e0-98ee-ae8114c18bda {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.172329] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dabbe03-18a8-4e55-bfab-08c73468abd7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.197504] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7da1d0f4-5dbe-4243-ac87-e476d5870882 could not be found. [ 651.197768] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 651.197956] env[63021]: INFO nova.compute.manager [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Took 0.04 seconds to destroy the instance on the hypervisor. [ 651.198240] env[63021]: DEBUG oslo.service.loopingcall [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 651.198521] env[63021]: DEBUG nova.compute.manager [-] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 651.198597] env[63021]: DEBUG nova.network.neutron [-] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 651.221020] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Acquiring lock "d02adae0-724e-4dbe-a6c5-d476805d3d57" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.221276] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Lock "d02adae0-724e-4dbe-a6c5-d476805d3d57" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.239108] env[63021]: DEBUG nova.network.neutron [-] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 651.330788] env[63021]: DEBUG nova.network.neutron [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Successfully created port: ed117e94-79ec-45ad-9605-16a99e4fb593 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 651.390682] env[63021]: DEBUG nova.network.neutron [-] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.524107] env[63021]: DEBUG nova.compute.manager [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 651.682569] env[63021]: DEBUG nova.network.neutron [req-c8118252-8a68-4436-9716-dc10fefc992f req-95615165-a312-411b-ab80-32362db4d702 service nova] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 651.741562] env[63021]: DEBUG nova.network.neutron [-] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.891835] env[63021]: DEBUG nova.network.neutron [req-c8118252-8a68-4436-9716-dc10fefc992f req-95615165-a312-411b-ab80-32362db4d702 service nova] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.892960] env[63021]: INFO nova.compute.manager [-] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Took 1.02 seconds to deallocate network for instance. [ 651.895060] env[63021]: DEBUG nova.compute.claims [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 651.895247] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.941248] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46a100e4-1f0b-458e-9f06-01bd87fde0f8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.950228] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12f7d86b-6146-4e98-be7d-4f9b5fb63ca2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.985520] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4f2f4af-dd69-4fe0-b16e-07674c5c6bc2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.993228] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bea27d7-bc48-44ee-b40f-2e199a6dfd9f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.006382] env[63021]: DEBUG nova.compute.provider_tree [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 652.145994] env[63021]: DEBUG oslo_concurrency.lockutils [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Acquiring lock "f77da305-9cc9-415d-a808-c2136212c456" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.146484] env[63021]: DEBUG oslo_concurrency.lockutils [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Lock "f77da305-9cc9-415d-a808-c2136212c456" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.246718] env[63021]: INFO nova.compute.manager [-] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Took 1.05 seconds to deallocate network for instance. [ 652.251266] env[63021]: DEBUG nova.compute.claims [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 652.251534] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.279698] env[63021]: ERROR nova.compute.manager [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ed117e94-79ec-45ad-9605-16a99e4fb593, please check neutron logs for more information. [ 652.279698] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 652.279698] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.279698] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 652.279698] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 652.279698] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 652.279698] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 652.279698] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 652.279698] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.279698] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 652.279698] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.279698] env[63021]: ERROR nova.compute.manager raise self.value [ 652.279698] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 652.279698] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 652.279698] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.279698] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 652.280134] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.280134] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 652.280134] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ed117e94-79ec-45ad-9605-16a99e4fb593, please check neutron logs for more information. [ 652.280134] env[63021]: ERROR nova.compute.manager [ 652.280236] env[63021]: Traceback (most recent call last): [ 652.280272] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 652.280272] env[63021]: listener.cb(fileno) [ 652.280272] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 652.280272] env[63021]: result = function(*args, **kwargs) [ 652.280272] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 652.280272] env[63021]: return func(*args, **kwargs) [ 652.280463] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 652.280463] env[63021]: raise e [ 652.280463] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.280463] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 652.280463] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 652.280463] env[63021]: created_port_ids = self._update_ports_for_instance( [ 652.280463] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 652.280463] env[63021]: with excutils.save_and_reraise_exception(): [ 652.280463] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.280463] env[63021]: self.force_reraise() [ 652.280463] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.280463] env[63021]: raise self.value [ 652.280463] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 652.280463] env[63021]: updated_port = self._update_port( [ 652.280463] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.280463] env[63021]: _ensure_no_port_binding_failure(port) [ 652.280463] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.280463] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 652.280463] env[63021]: nova.exception.PortBindingFailed: Binding failed for port ed117e94-79ec-45ad-9605-16a99e4fb593, please check neutron logs for more information. [ 652.280463] env[63021]: Removing descriptor: 18 [ 652.396083] env[63021]: DEBUG oslo_concurrency.lockutils [req-c8118252-8a68-4436-9716-dc10fefc992f req-95615165-a312-411b-ab80-32362db4d702 service nova] Releasing lock "refresh_cache-7da1d0f4-5dbe-4243-ac87-e476d5870882" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.497583] env[63021]: DEBUG nova.compute.manager [req-5958780d-9b9d-4ad8-a78d-b51618e3aaed req-f5c1471c-839b-4ee1-9bea-94cffd52185e service nova] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Received event network-vif-deleted-0b0dcc76-2a8c-4f06-b32e-22e39f072235 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 652.497788] env[63021]: DEBUG nova.compute.manager [req-5958780d-9b9d-4ad8-a78d-b51618e3aaed req-f5c1471c-839b-4ee1-9bea-94cffd52185e service nova] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Received event network-changed-ed117e94-79ec-45ad-9605-16a99e4fb593 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 652.497995] env[63021]: DEBUG nova.compute.manager [req-5958780d-9b9d-4ad8-a78d-b51618e3aaed req-f5c1471c-839b-4ee1-9bea-94cffd52185e service nova] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Refreshing instance network info cache due to event network-changed-ed117e94-79ec-45ad-9605-16a99e4fb593. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 652.498508] env[63021]: DEBUG oslo_concurrency.lockutils [req-5958780d-9b9d-4ad8-a78d-b51618e3aaed req-f5c1471c-839b-4ee1-9bea-94cffd52185e service nova] Acquiring lock "refresh_cache-c76dfad0-2c67-4fe4-8c86-04ece5c80ea8" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 652.498508] env[63021]: DEBUG oslo_concurrency.lockutils [req-5958780d-9b9d-4ad8-a78d-b51618e3aaed req-f5c1471c-839b-4ee1-9bea-94cffd52185e service nova] Acquired lock "refresh_cache-c76dfad0-2c67-4fe4-8c86-04ece5c80ea8" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.498817] env[63021]: DEBUG nova.network.neutron [req-5958780d-9b9d-4ad8-a78d-b51618e3aaed req-f5c1471c-839b-4ee1-9bea-94cffd52185e service nova] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Refreshing network info cache for port ed117e94-79ec-45ad-9605-16a99e4fb593 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 652.509016] env[63021]: DEBUG nova.scheduler.client.report [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 652.536253] env[63021]: DEBUG nova.compute.manager [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 652.562580] env[63021]: DEBUG nova.virt.hardware [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 652.563147] env[63021]: DEBUG nova.virt.hardware [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 652.563147] env[63021]: DEBUG nova.virt.hardware [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 652.563147] env[63021]: DEBUG nova.virt.hardware [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 652.563425] env[63021]: DEBUG nova.virt.hardware [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 652.563425] env[63021]: DEBUG nova.virt.hardware [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 652.563661] env[63021]: DEBUG nova.virt.hardware [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 652.563815] env[63021]: DEBUG nova.virt.hardware [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 652.563978] env[63021]: DEBUG nova.virt.hardware [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 652.564150] env[63021]: DEBUG nova.virt.hardware [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 652.564383] env[63021]: DEBUG nova.virt.hardware [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 652.565443] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6370da08-c853-4a57-a479-70d2b7645c7f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.575091] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92464a6d-5ccc-4399-9d3f-2e62d8535c0d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.588817] env[63021]: ERROR nova.compute.manager [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ed117e94-79ec-45ad-9605-16a99e4fb593, please check neutron logs for more information. [ 652.588817] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Traceback (most recent call last): [ 652.588817] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 652.588817] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] yield resources [ 652.588817] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 652.588817] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] self.driver.spawn(context, instance, image_meta, [ 652.588817] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 652.588817] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 652.588817] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 652.588817] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] vm_ref = self.build_virtual_machine(instance, [ 652.588817] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 652.590174] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] vif_infos = vmwarevif.get_vif_info(self._session, [ 652.590174] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 652.590174] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] for vif in network_info: [ 652.590174] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 652.590174] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] return self._sync_wrapper(fn, *args, **kwargs) [ 652.590174] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 652.590174] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] self.wait() [ 652.590174] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 652.590174] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] self[:] = self._gt.wait() [ 652.590174] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 652.590174] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] return self._exit_event.wait() [ 652.590174] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 652.590174] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] current.throw(*self._exc) [ 652.590824] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 652.590824] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] result = function(*args, **kwargs) [ 652.590824] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 652.590824] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] return func(*args, **kwargs) [ 652.590824] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 652.590824] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] raise e [ 652.590824] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.590824] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] nwinfo = self.network_api.allocate_for_instance( [ 652.590824] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 652.590824] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] created_port_ids = self._update_ports_for_instance( [ 652.590824] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 652.590824] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] with excutils.save_and_reraise_exception(): [ 652.590824] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.591298] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] self.force_reraise() [ 652.591298] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.591298] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] raise self.value [ 652.591298] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 652.591298] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] updated_port = self._update_port( [ 652.591298] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.591298] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] _ensure_no_port_binding_failure(port) [ 652.591298] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.591298] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] raise exception.PortBindingFailed(port_id=port['id']) [ 652.591298] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] nova.exception.PortBindingFailed: Binding failed for port ed117e94-79ec-45ad-9605-16a99e4fb593, please check neutron logs for more information. [ 652.591298] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] [ 652.591298] env[63021]: INFO nova.compute.manager [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Terminating instance [ 652.591657] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Acquiring lock "refresh_cache-c76dfad0-2c67-4fe4-8c86-04ece5c80ea8" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.014074] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.504s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 653.014738] env[63021]: DEBUG nova.compute.manager [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 653.018031] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 18.641s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 653.018303] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 653.018650] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63021) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 653.018996] env[63021]: DEBUG oslo_concurrency.lockutils [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.540s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 653.020563] env[63021]: INFO nova.compute.claims [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 653.024294] env[63021]: DEBUG nova.network.neutron [req-5958780d-9b9d-4ad8-a78d-b51618e3aaed req-f5c1471c-839b-4ee1-9bea-94cffd52185e service nova] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.026839] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a645f8e7-352d-49d6-814d-315a07bce1c3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.037097] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e9e11ce-8dd4-49b4-9b20-bcf1d234c53f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.054998] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2a24fe9-2f45-4318-878d-5a64742a162b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.063668] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24046eb0-4843-4b43-841d-204d195fb574 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.097576] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181310MB free_disk=139GB free_vcpus=48 pci_devices=None {{(pid=63021) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 653.097669] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 653.153571] env[63021]: DEBUG nova.network.neutron [req-5958780d-9b9d-4ad8-a78d-b51618e3aaed req-f5c1471c-839b-4ee1-9bea-94cffd52185e service nova] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.527677] env[63021]: DEBUG nova.compute.utils [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 653.529147] env[63021]: DEBUG nova.compute.manager [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 653.529374] env[63021]: DEBUG nova.network.neutron [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 653.610532] env[63021]: DEBUG nova.policy [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f37ac2ae39ec43669736e1de99adb298', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f64fe0a46b07403f9ce4317b8a743121', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 653.656831] env[63021]: DEBUG oslo_concurrency.lockutils [req-5958780d-9b9d-4ad8-a78d-b51618e3aaed req-f5c1471c-839b-4ee1-9bea-94cffd52185e service nova] Releasing lock "refresh_cache-c76dfad0-2c67-4fe4-8c86-04ece5c80ea8" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 653.657149] env[63021]: DEBUG nova.compute.manager [req-5958780d-9b9d-4ad8-a78d-b51618e3aaed req-f5c1471c-839b-4ee1-9bea-94cffd52185e service nova] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Received event network-vif-deleted-ed117e94-79ec-45ad-9605-16a99e4fb593 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 653.657516] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Acquired lock "refresh_cache-c76dfad0-2c67-4fe4-8c86-04ece5c80ea8" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.657688] env[63021]: DEBUG nova.network.neutron [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 654.023745] env[63021]: DEBUG nova.network.neutron [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Successfully created port: 2f32ab2a-287e-403b-b713-052e29aea30e {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 654.033887] env[63021]: DEBUG nova.compute.manager [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 654.175848] env[63021]: DEBUG nova.network.neutron [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 654.289598] env[63021]: DEBUG nova.network.neutron [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.452333] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-453337fa-d93e-4493-927f-47f1248a0d48 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.461017] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec7ebd0a-9cf4-4a61-9043-a221d87f8e79 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.493981] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82da643d-b742-493e-8fa4-d7b8675f2142 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.501569] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81b160ab-d83f-4272-a8d0-277d8504aeb1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.515089] env[63021]: DEBUG nova.compute.provider_tree [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 654.799022] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Releasing lock "refresh_cache-c76dfad0-2c67-4fe4-8c86-04ece5c80ea8" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 654.799022] env[63021]: DEBUG nova.compute.manager [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 654.799022] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 654.799022] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5bf6e72a-9c06-4097-9815-fb4b09b448d8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.806054] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d235fd1-6a95-4dcf-8171-f0baffbfe4fb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.833489] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c76dfad0-2c67-4fe4-8c86-04ece5c80ea8 could not be found. [ 654.833489] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 654.833489] env[63021]: INFO nova.compute.manager [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Took 0.04 seconds to destroy the instance on the hypervisor. [ 654.833489] env[63021]: DEBUG oslo.service.loopingcall [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 654.833489] env[63021]: DEBUG nova.compute.manager [-] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 654.833489] env[63021]: DEBUG nova.network.neutron [-] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 654.851426] env[63021]: DEBUG nova.network.neutron [-] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.021393] env[63021]: DEBUG nova.scheduler.client.report [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 655.050565] env[63021]: DEBUG nova.compute.manager [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 655.085239] env[63021]: DEBUG nova.virt.hardware [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 655.085542] env[63021]: DEBUG nova.virt.hardware [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 655.085740] env[63021]: DEBUG nova.virt.hardware [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 655.085964] env[63021]: DEBUG nova.virt.hardware [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 655.086580] env[63021]: DEBUG nova.virt.hardware [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 655.086580] env[63021]: DEBUG nova.virt.hardware [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 655.086580] env[63021]: DEBUG nova.virt.hardware [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 655.086736] env[63021]: DEBUG nova.virt.hardware [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 655.086954] env[63021]: DEBUG nova.virt.hardware [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 655.087164] env[63021]: DEBUG nova.virt.hardware [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 655.087369] env[63021]: DEBUG nova.virt.hardware [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 655.088250] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8a30adb-b076-44f1-91f2-a564336eb64a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.096791] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b407e5f-189a-403e-9b13-338ee135eb1d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.225506] env[63021]: DEBUG nova.compute.manager [req-0b017fd0-11c3-42b9-ac14-7310959ea79d req-d0920735-9408-4c93-a78b-8e33f4d3aaf8 service nova] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Received event network-changed-2f32ab2a-287e-403b-b713-052e29aea30e {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 655.225775] env[63021]: DEBUG nova.compute.manager [req-0b017fd0-11c3-42b9-ac14-7310959ea79d req-d0920735-9408-4c93-a78b-8e33f4d3aaf8 service nova] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Refreshing instance network info cache due to event network-changed-2f32ab2a-287e-403b-b713-052e29aea30e. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 655.225970] env[63021]: DEBUG oslo_concurrency.lockutils [req-0b017fd0-11c3-42b9-ac14-7310959ea79d req-d0920735-9408-4c93-a78b-8e33f4d3aaf8 service nova] Acquiring lock "refresh_cache-1aaa84d5-c5d5-4321-a98f-da261de5d364" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 655.226139] env[63021]: DEBUG oslo_concurrency.lockutils [req-0b017fd0-11c3-42b9-ac14-7310959ea79d req-d0920735-9408-4c93-a78b-8e33f4d3aaf8 service nova] Acquired lock "refresh_cache-1aaa84d5-c5d5-4321-a98f-da261de5d364" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.226302] env[63021]: DEBUG nova.network.neutron [req-0b017fd0-11c3-42b9-ac14-7310959ea79d req-d0920735-9408-4c93-a78b-8e33f4d3aaf8 service nova] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Refreshing network info cache for port 2f32ab2a-287e-403b-b713-052e29aea30e {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 655.354548] env[63021]: DEBUG nova.network.neutron [-] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.404422] env[63021]: ERROR nova.compute.manager [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2f32ab2a-287e-403b-b713-052e29aea30e, please check neutron logs for more information. [ 655.404422] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 655.404422] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 655.404422] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 655.404422] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 655.404422] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 655.404422] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 655.404422] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 655.404422] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 655.404422] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 655.404422] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 655.404422] env[63021]: ERROR nova.compute.manager raise self.value [ 655.404422] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 655.404422] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 655.404422] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 655.404422] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 655.404870] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 655.404870] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 655.404870] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2f32ab2a-287e-403b-b713-052e29aea30e, please check neutron logs for more information. [ 655.404870] env[63021]: ERROR nova.compute.manager [ 655.404870] env[63021]: Traceback (most recent call last): [ 655.404870] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 655.404870] env[63021]: listener.cb(fileno) [ 655.404870] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 655.404870] env[63021]: result = function(*args, **kwargs) [ 655.404870] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 655.404870] env[63021]: return func(*args, **kwargs) [ 655.404870] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 655.404870] env[63021]: raise e [ 655.404870] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 655.404870] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 655.404870] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 655.404870] env[63021]: created_port_ids = self._update_ports_for_instance( [ 655.404870] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 655.404870] env[63021]: with excutils.save_and_reraise_exception(): [ 655.404870] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 655.404870] env[63021]: self.force_reraise() [ 655.404870] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 655.404870] env[63021]: raise self.value [ 655.404870] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 655.404870] env[63021]: updated_port = self._update_port( [ 655.404870] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 655.404870] env[63021]: _ensure_no_port_binding_failure(port) [ 655.404870] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 655.404870] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 655.405592] env[63021]: nova.exception.PortBindingFailed: Binding failed for port 2f32ab2a-287e-403b-b713-052e29aea30e, please check neutron logs for more information. [ 655.405592] env[63021]: Removing descriptor: 18 [ 655.405592] env[63021]: ERROR nova.compute.manager [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2f32ab2a-287e-403b-b713-052e29aea30e, please check neutron logs for more information. [ 655.405592] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Traceback (most recent call last): [ 655.405592] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 655.405592] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] yield resources [ 655.405592] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 655.405592] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] self.driver.spawn(context, instance, image_meta, [ 655.405592] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 655.405592] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] self._vmops.spawn(context, instance, image_meta, injected_files, [ 655.405592] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 655.405592] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] vm_ref = self.build_virtual_machine(instance, [ 655.405897] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 655.405897] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] vif_infos = vmwarevif.get_vif_info(self._session, [ 655.405897] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 655.405897] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] for vif in network_info: [ 655.405897] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 655.405897] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] return self._sync_wrapper(fn, *args, **kwargs) [ 655.405897] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 655.405897] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] self.wait() [ 655.405897] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 655.405897] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] self[:] = self._gt.wait() [ 655.405897] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 655.405897] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] return self._exit_event.wait() [ 655.405897] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 655.406442] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] result = hub.switch() [ 655.406442] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 655.406442] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] return self.greenlet.switch() [ 655.406442] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 655.406442] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] result = function(*args, **kwargs) [ 655.406442] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 655.406442] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] return func(*args, **kwargs) [ 655.406442] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 655.406442] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] raise e [ 655.406442] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 655.406442] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] nwinfo = self.network_api.allocate_for_instance( [ 655.406442] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 655.406442] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] created_port_ids = self._update_ports_for_instance( [ 655.406769] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 655.406769] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] with excutils.save_and_reraise_exception(): [ 655.406769] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 655.406769] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] self.force_reraise() [ 655.406769] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 655.406769] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] raise self.value [ 655.406769] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 655.406769] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] updated_port = self._update_port( [ 655.406769] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 655.406769] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] _ensure_no_port_binding_failure(port) [ 655.406769] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 655.406769] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] raise exception.PortBindingFailed(port_id=port['id']) [ 655.407701] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] nova.exception.PortBindingFailed: Binding failed for port 2f32ab2a-287e-403b-b713-052e29aea30e, please check neutron logs for more information. [ 655.407701] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] [ 655.407701] env[63021]: INFO nova.compute.manager [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Terminating instance [ 655.407701] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Acquiring lock "refresh_cache-1aaa84d5-c5d5-4321-a98f-da261de5d364" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 655.531021] env[63021]: DEBUG oslo_concurrency.lockutils [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.509s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 655.531021] env[63021]: DEBUG nova.compute.manager [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 655.531394] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.726s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 655.752499] env[63021]: DEBUG nova.network.neutron [req-0b017fd0-11c3-42b9-ac14-7310959ea79d req-d0920735-9408-4c93-a78b-8e33f4d3aaf8 service nova] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.830538] env[63021]: DEBUG nova.network.neutron [req-0b017fd0-11c3-42b9-ac14-7310959ea79d req-d0920735-9408-4c93-a78b-8e33f4d3aaf8 service nova] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.862050] env[63021]: INFO nova.compute.manager [-] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Took 1.03 seconds to deallocate network for instance. [ 655.862886] env[63021]: DEBUG nova.compute.claims [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 655.863175] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.037511] env[63021]: DEBUG nova.compute.utils [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 656.041972] env[63021]: DEBUG nova.compute.manager [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 656.042171] env[63021]: DEBUG nova.network.neutron [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 656.128721] env[63021]: DEBUG nova.policy [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e063a962e064461994036bd68d51be76', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3275d10f82e94ecda609f9c7774146f5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 656.333433] env[63021]: DEBUG oslo_concurrency.lockutils [req-0b017fd0-11c3-42b9-ac14-7310959ea79d req-d0920735-9408-4c93-a78b-8e33f4d3aaf8 service nova] Releasing lock "refresh_cache-1aaa84d5-c5d5-4321-a98f-da261de5d364" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 656.333853] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Acquired lock "refresh_cache-1aaa84d5-c5d5-4321-a98f-da261de5d364" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.334051] env[63021]: DEBUG nova.network.neutron [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 656.449943] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aa32698-6c4f-4337-8b5a-bd63d4e0f9bd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.454285] env[63021]: DEBUG nova.network.neutron [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Successfully created port: 45c257ff-e39a-4ca8-88b1-bc84407ea2d8 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 656.461490] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-229be6ec-18c9-4ccb-87ef-92eee5f5ac35 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.497897] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd7220da-a948-4f07-b17e-2fcbd38423ef {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.506436] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e66b0a7d-623c-49ee-a362-1b68b5532304 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.522272] env[63021]: DEBUG nova.compute.provider_tree [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 656.542334] env[63021]: DEBUG nova.compute.manager [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 656.853078] env[63021]: DEBUG nova.network.neutron [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 657.026356] env[63021]: DEBUG nova.scheduler.client.report [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 657.039234] env[63021]: DEBUG nova.network.neutron [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.329238] env[63021]: DEBUG nova.compute.manager [req-82acdaff-b693-40a3-8160-5be324a78721 req-a52da3b4-90e2-43fb-9612-afea81462645 service nova] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Received event network-vif-deleted-2f32ab2a-287e-403b-b713-052e29aea30e {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 657.439228] env[63021]: ERROR nova.compute.manager [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 45c257ff-e39a-4ca8-88b1-bc84407ea2d8, please check neutron logs for more information. [ 657.439228] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 657.439228] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.439228] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 657.439228] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 657.439228] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 657.439228] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 657.439228] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 657.439228] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.439228] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 657.439228] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.439228] env[63021]: ERROR nova.compute.manager raise self.value [ 657.439228] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 657.439228] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 657.439228] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.439228] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 657.439631] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.439631] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 657.439631] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 45c257ff-e39a-4ca8-88b1-bc84407ea2d8, please check neutron logs for more information. [ 657.439631] env[63021]: ERROR nova.compute.manager [ 657.439631] env[63021]: Traceback (most recent call last): [ 657.439631] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 657.439631] env[63021]: listener.cb(fileno) [ 657.439631] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 657.439631] env[63021]: result = function(*args, **kwargs) [ 657.439631] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 657.439631] env[63021]: return func(*args, **kwargs) [ 657.439631] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 657.439631] env[63021]: raise e [ 657.439631] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.439631] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 657.439631] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 657.439631] env[63021]: created_port_ids = self._update_ports_for_instance( [ 657.439631] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 657.439631] env[63021]: with excutils.save_and_reraise_exception(): [ 657.439631] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.439631] env[63021]: self.force_reraise() [ 657.439631] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.439631] env[63021]: raise self.value [ 657.439631] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 657.439631] env[63021]: updated_port = self._update_port( [ 657.439631] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.439631] env[63021]: _ensure_no_port_binding_failure(port) [ 657.439631] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.439631] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 657.440383] env[63021]: nova.exception.PortBindingFailed: Binding failed for port 45c257ff-e39a-4ca8-88b1-bc84407ea2d8, please check neutron logs for more information. [ 657.440383] env[63021]: Removing descriptor: 18 [ 657.537016] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.537016] env[63021]: ERROR nova.compute.manager [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port af20fd90-4fdd-4898-9acf-ee54c4d9f27b, please check neutron logs for more information. [ 657.537016] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Traceback (most recent call last): [ 657.537016] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 657.537016] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] self.driver.spawn(context, instance, image_meta, [ 657.537016] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 657.537016] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] self._vmops.spawn(context, instance, image_meta, injected_files, [ 657.537016] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 657.537016] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] vm_ref = self.build_virtual_machine(instance, [ 657.537332] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 657.537332] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] vif_infos = vmwarevif.get_vif_info(self._session, [ 657.537332] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 657.537332] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] for vif in network_info: [ 657.537332] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 657.537332] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] return self._sync_wrapper(fn, *args, **kwargs) [ 657.537332] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 657.537332] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] self.wait() [ 657.537332] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 657.537332] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] self[:] = self._gt.wait() [ 657.537332] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 657.537332] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] return self._exit_event.wait() [ 657.537332] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 657.537632] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] result = hub.switch() [ 657.537632] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 657.537632] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] return self.greenlet.switch() [ 657.537632] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 657.537632] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] result = function(*args, **kwargs) [ 657.537632] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 657.537632] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] return func(*args, **kwargs) [ 657.537632] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 657.537632] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] raise e [ 657.537632] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.537632] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] nwinfo = self.network_api.allocate_for_instance( [ 657.537632] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 657.537632] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] created_port_ids = self._update_ports_for_instance( [ 657.537922] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 657.537922] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] with excutils.save_and_reraise_exception(): [ 657.537922] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.537922] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] self.force_reraise() [ 657.537922] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.537922] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] raise self.value [ 657.537922] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 657.537922] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] updated_port = self._update_port( [ 657.537922] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.537922] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] _ensure_no_port_binding_failure(port) [ 657.537922] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.537922] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] raise exception.PortBindingFailed(port_id=port['id']) [ 657.538209] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] nova.exception.PortBindingFailed: Binding failed for port af20fd90-4fdd-4898-9acf-ee54c4d9f27b, please check neutron logs for more information. [ 657.538209] env[63021]: ERROR nova.compute.manager [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] [ 657.538209] env[63021]: DEBUG nova.compute.utils [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Binding failed for port af20fd90-4fdd-4898-9acf-ee54c4d9f27b, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 657.538209] env[63021]: DEBUG oslo_concurrency.lockutils [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.033s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.543310] env[63021]: INFO nova.compute.claims [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 657.545887] env[63021]: DEBUG nova.compute.manager [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Build of instance ab11dead-3fe6-46d3-92e1-cdca84ebe508 was re-scheduled: Binding failed for port af20fd90-4fdd-4898-9acf-ee54c4d9f27b, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 657.546427] env[63021]: DEBUG nova.compute.manager [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 657.546696] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Acquiring lock "refresh_cache-ab11dead-3fe6-46d3-92e1-cdca84ebe508" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.546882] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Acquired lock "refresh_cache-ab11dead-3fe6-46d3-92e1-cdca84ebe508" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.547088] env[63021]: DEBUG nova.network.neutron [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 657.548959] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Releasing lock "refresh_cache-1aaa84d5-c5d5-4321-a98f-da261de5d364" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.550080] env[63021]: DEBUG nova.compute.manager [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 657.550080] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 657.550434] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1d91164c-c0b0-4293-9bd6-65eabd7eff3e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.554112] env[63021]: DEBUG nova.compute.manager [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 657.563623] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71c73669-c8ec-46cb-8ab3-d65652b7958a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.586372] env[63021]: DEBUG nova.virt.hardware [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:44:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='c5c88a1f-b85b-4c94-82e7-dbd1e5698cca',id=31,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1744080769',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 657.586787] env[63021]: DEBUG nova.virt.hardware [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 657.586986] env[63021]: DEBUG nova.virt.hardware [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 657.587194] env[63021]: DEBUG nova.virt.hardware [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 657.587346] env[63021]: DEBUG nova.virt.hardware [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 657.587526] env[63021]: DEBUG nova.virt.hardware [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 657.587765] env[63021]: DEBUG nova.virt.hardware [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 657.587933] env[63021]: DEBUG nova.virt.hardware [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 657.588131] env[63021]: DEBUG nova.virt.hardware [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 657.588762] env[63021]: DEBUG nova.virt.hardware [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 657.588980] env[63021]: DEBUG nova.virt.hardware [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 657.590616] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e62258e0-43f6-4e65-a8a7-e78a9ae4bb59 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.599099] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1aaa84d5-c5d5-4321-a98f-da261de5d364 could not be found. [ 657.599312] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 657.599491] env[63021]: INFO nova.compute.manager [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Took 0.05 seconds to destroy the instance on the hypervisor. [ 657.599732] env[63021]: DEBUG oslo.service.loopingcall [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 657.600987] env[63021]: DEBUG nova.compute.manager [-] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 657.601098] env[63021]: DEBUG nova.network.neutron [-] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 657.606434] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1d94125-8ea9-4618-b1ae-c5b55f6371e9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.620763] env[63021]: ERROR nova.compute.manager [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 45c257ff-e39a-4ca8-88b1-bc84407ea2d8, please check neutron logs for more information. [ 657.620763] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Traceback (most recent call last): [ 657.620763] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 657.620763] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] yield resources [ 657.620763] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 657.620763] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] self.driver.spawn(context, instance, image_meta, [ 657.620763] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 657.620763] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 657.620763] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 657.620763] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] vm_ref = self.build_virtual_machine(instance, [ 657.620763] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 657.621145] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] vif_infos = vmwarevif.get_vif_info(self._session, [ 657.621145] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 657.621145] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] for vif in network_info: [ 657.621145] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 657.621145] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] return self._sync_wrapper(fn, *args, **kwargs) [ 657.621145] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 657.621145] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] self.wait() [ 657.621145] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 657.621145] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] self[:] = self._gt.wait() [ 657.621145] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 657.621145] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] return self._exit_event.wait() [ 657.621145] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 657.621145] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] current.throw(*self._exc) [ 657.621529] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 657.621529] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] result = function(*args, **kwargs) [ 657.621529] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 657.621529] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] return func(*args, **kwargs) [ 657.621529] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 657.621529] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] raise e [ 657.621529] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.621529] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] nwinfo = self.network_api.allocate_for_instance( [ 657.621529] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 657.621529] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] created_port_ids = self._update_ports_for_instance( [ 657.621529] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 657.621529] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] with excutils.save_and_reraise_exception(): [ 657.621529] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.621898] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] self.force_reraise() [ 657.621898] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.621898] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] raise self.value [ 657.621898] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 657.621898] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] updated_port = self._update_port( [ 657.621898] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.621898] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] _ensure_no_port_binding_failure(port) [ 657.621898] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.621898] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] raise exception.PortBindingFailed(port_id=port['id']) [ 657.621898] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] nova.exception.PortBindingFailed: Binding failed for port 45c257ff-e39a-4ca8-88b1-bc84407ea2d8, please check neutron logs for more information. [ 657.621898] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] [ 657.621898] env[63021]: INFO nova.compute.manager [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Terminating instance [ 657.623009] env[63021]: DEBUG nova.network.neutron [-] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 657.624249] env[63021]: DEBUG oslo_concurrency.lockutils [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Acquiring lock "refresh_cache-f5f546f8-291f-430c-95fb-c78507b0087f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.624427] env[63021]: DEBUG oslo_concurrency.lockutils [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Acquired lock "refresh_cache-f5f546f8-291f-430c-95fb-c78507b0087f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.624560] env[63021]: DEBUG nova.network.neutron [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 658.075959] env[63021]: DEBUG nova.network.neutron [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.125917] env[63021]: DEBUG nova.network.neutron [-] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.148208] env[63021]: DEBUG nova.network.neutron [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.168030] env[63021]: DEBUG nova.network.neutron [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.237199] env[63021]: DEBUG nova.network.neutron [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.630040] env[63021]: INFO nova.compute.manager [-] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Took 1.03 seconds to deallocate network for instance. [ 658.635850] env[63021]: DEBUG nova.compute.claims [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 658.636029] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 658.674222] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Releasing lock "refresh_cache-ab11dead-3fe6-46d3-92e1-cdca84ebe508" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 658.674558] env[63021]: DEBUG nova.compute.manager [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 658.674822] env[63021]: DEBUG nova.compute.manager [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 658.674999] env[63021]: DEBUG nova.network.neutron [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 658.690275] env[63021]: DEBUG nova.network.neutron [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.741969] env[63021]: DEBUG oslo_concurrency.lockutils [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Releasing lock "refresh_cache-f5f546f8-291f-430c-95fb-c78507b0087f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 658.742377] env[63021]: DEBUG nova.compute.manager [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 658.742598] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 658.743085] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-07d52c97-cc2a-4172-9c53-4804613cfcf7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.753617] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1b21083-9fe4-4899-a2ad-ee0c20a0df5c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.783026] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f5f546f8-291f-430c-95fb-c78507b0087f could not be found. [ 658.783377] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 658.783468] env[63021]: INFO nova.compute.manager [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 658.783734] env[63021]: DEBUG oslo.service.loopingcall [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 658.786323] env[63021]: DEBUG nova.compute.manager [-] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 658.786479] env[63021]: DEBUG nova.network.neutron [-] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 658.827260] env[63021]: DEBUG nova.network.neutron [-] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 659.072483] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8fa5023-9c46-40c0-af12-575c58c9bb4c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.082948] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbea7a57-2f18-4dad-a458-5da07e530b75 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.123243] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88bf62d8-b903-4421-9e58-4c5aeadb982c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.130944] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f49ac8c4-ec74-4f9f-a8ff-ae29c9352f54 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.146957] env[63021]: DEBUG nova.compute.provider_tree [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 659.193600] env[63021]: DEBUG nova.network.neutron [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.329731] env[63021]: DEBUG nova.network.neutron [-] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.352509] env[63021]: DEBUG nova.compute.manager [req-93c17f1b-56fa-4587-b620-40f7556367e8 req-273078a0-8e8e-4e80-be66-99d757c7ea78 service nova] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Received event network-changed-45c257ff-e39a-4ca8-88b1-bc84407ea2d8 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 659.352929] env[63021]: DEBUG nova.compute.manager [req-93c17f1b-56fa-4587-b620-40f7556367e8 req-273078a0-8e8e-4e80-be66-99d757c7ea78 service nova] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Refreshing instance network info cache due to event network-changed-45c257ff-e39a-4ca8-88b1-bc84407ea2d8. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 659.354024] env[63021]: DEBUG oslo_concurrency.lockutils [req-93c17f1b-56fa-4587-b620-40f7556367e8 req-273078a0-8e8e-4e80-be66-99d757c7ea78 service nova] Acquiring lock "refresh_cache-f5f546f8-291f-430c-95fb-c78507b0087f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 659.354024] env[63021]: DEBUG oslo_concurrency.lockutils [req-93c17f1b-56fa-4587-b620-40f7556367e8 req-273078a0-8e8e-4e80-be66-99d757c7ea78 service nova] Acquired lock "refresh_cache-f5f546f8-291f-430c-95fb-c78507b0087f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.354024] env[63021]: DEBUG nova.network.neutron [req-93c17f1b-56fa-4587-b620-40f7556367e8 req-273078a0-8e8e-4e80-be66-99d757c7ea78 service nova] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Refreshing network info cache for port 45c257ff-e39a-4ca8-88b1-bc84407ea2d8 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 659.650291] env[63021]: DEBUG nova.scheduler.client.report [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 659.698252] env[63021]: INFO nova.compute.manager [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] [instance: ab11dead-3fe6-46d3-92e1-cdca84ebe508] Took 1.02 seconds to deallocate network for instance. [ 659.833489] env[63021]: INFO nova.compute.manager [-] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Took 1.05 seconds to deallocate network for instance. [ 659.836856] env[63021]: DEBUG nova.compute.claims [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 659.837062] env[63021]: DEBUG oslo_concurrency.lockutils [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 659.889866] env[63021]: DEBUG nova.network.neutron [req-93c17f1b-56fa-4587-b620-40f7556367e8 req-273078a0-8e8e-4e80-be66-99d757c7ea78 service nova] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 659.984022] env[63021]: DEBUG nova.network.neutron [req-93c17f1b-56fa-4587-b620-40f7556367e8 req-273078a0-8e8e-4e80-be66-99d757c7ea78 service nova] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.158029] env[63021]: DEBUG oslo_concurrency.lockutils [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.620s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.158622] env[63021]: DEBUG nova.compute.manager [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 660.161385] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.647s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.163417] env[63021]: INFO nova.compute.claims [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 660.487606] env[63021]: DEBUG oslo_concurrency.lockutils [req-93c17f1b-56fa-4587-b620-40f7556367e8 req-273078a0-8e8e-4e80-be66-99d757c7ea78 service nova] Releasing lock "refresh_cache-f5f546f8-291f-430c-95fb-c78507b0087f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 660.487891] env[63021]: DEBUG nova.compute.manager [req-93c17f1b-56fa-4587-b620-40f7556367e8 req-273078a0-8e8e-4e80-be66-99d757c7ea78 service nova] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Received event network-vif-deleted-45c257ff-e39a-4ca8-88b1-bc84407ea2d8 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 660.668265] env[63021]: DEBUG nova.compute.utils [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 660.671921] env[63021]: DEBUG nova.compute.manager [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 660.672196] env[63021]: DEBUG nova.network.neutron [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 660.730612] env[63021]: INFO nova.scheduler.client.report [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Deleted allocations for instance ab11dead-3fe6-46d3-92e1-cdca84ebe508 [ 660.737090] env[63021]: DEBUG nova.policy [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0aa5f43b1e9c45419205078f65f64df4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '01adb33f517f49e5bccef38407e69a43', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 661.173191] env[63021]: DEBUG nova.compute.manager [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 661.183053] env[63021]: DEBUG nova.network.neutron [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Successfully created port: 31850b77-7688-482d-91fb-d74fdd83e293 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 661.243322] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b9a4f9f0-ec0a-4ba2-9eff-61a4bc8095bf tempest-ImagesOneServerTestJSON-1842864275 tempest-ImagesOneServerTestJSON-1842864275-project-member] Lock "ab11dead-3fe6-46d3-92e1-cdca84ebe508" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 93.647s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.539276] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f0ca35d-2e83-4728-8924-632c9d900434 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.545447] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2780b2e0-46d8-4c57-a269-65ccb15b98e6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.577578] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dab6eb3a-3878-4e87-9ae1-4f6e1d7404c2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.585978] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2db34b97-8a4a-4a24-bceb-ba0b8dbd54f6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.600645] env[63021]: DEBUG nova.compute.provider_tree [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 661.749248] env[63021]: DEBUG nova.compute.manager [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 662.103854] env[63021]: DEBUG nova.scheduler.client.report [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 662.188075] env[63021]: DEBUG nova.compute.manager [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 662.210447] env[63021]: DEBUG nova.virt.hardware [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 662.210692] env[63021]: DEBUG nova.virt.hardware [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 662.210846] env[63021]: DEBUG nova.virt.hardware [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 662.211037] env[63021]: DEBUG nova.virt.hardware [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 662.211186] env[63021]: DEBUG nova.virt.hardware [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 662.211362] env[63021]: DEBUG nova.virt.hardware [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 662.212207] env[63021]: DEBUG nova.virt.hardware [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 662.212207] env[63021]: DEBUG nova.virt.hardware [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 662.212207] env[63021]: DEBUG nova.virt.hardware [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 662.212207] env[63021]: DEBUG nova.virt.hardware [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 662.215680] env[63021]: DEBUG nova.virt.hardware [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 662.216582] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a720bd6-39f6-4e55-8a8a-80e6605410d1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.228422] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-999364ba-d531-460f-aa61-d64552a0de34 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.272230] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.608977] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.447s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 662.609555] env[63021]: DEBUG nova.compute.manager [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 662.612434] env[63021]: DEBUG oslo_concurrency.lockutils [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.936s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.616022] env[63021]: INFO nova.compute.claims [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 663.116586] env[63021]: DEBUG nova.compute.manager [req-3479c9eb-10e7-4710-9de5-ead7a8b8b91a req-beb926b8-732d-45d3-8831-6ba74e5d1a33 service nova] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Received event network-changed-31850b77-7688-482d-91fb-d74fdd83e293 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 663.116781] env[63021]: DEBUG nova.compute.manager [req-3479c9eb-10e7-4710-9de5-ead7a8b8b91a req-beb926b8-732d-45d3-8831-6ba74e5d1a33 service nova] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Refreshing instance network info cache due to event network-changed-31850b77-7688-482d-91fb-d74fdd83e293. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 663.116989] env[63021]: DEBUG oslo_concurrency.lockutils [req-3479c9eb-10e7-4710-9de5-ead7a8b8b91a req-beb926b8-732d-45d3-8831-6ba74e5d1a33 service nova] Acquiring lock "refresh_cache-05caabab-8c7d-4d27-8eb3-e7e1c2947aa6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.117143] env[63021]: DEBUG oslo_concurrency.lockutils [req-3479c9eb-10e7-4710-9de5-ead7a8b8b91a req-beb926b8-732d-45d3-8831-6ba74e5d1a33 service nova] Acquired lock "refresh_cache-05caabab-8c7d-4d27-8eb3-e7e1c2947aa6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.117301] env[63021]: DEBUG nova.network.neutron [req-3479c9eb-10e7-4710-9de5-ead7a8b8b91a req-beb926b8-732d-45d3-8831-6ba74e5d1a33 service nova] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Refreshing network info cache for port 31850b77-7688-482d-91fb-d74fdd83e293 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 663.120356] env[63021]: DEBUG nova.compute.utils [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 663.123733] env[63021]: DEBUG nova.compute.manager [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 663.123733] env[63021]: DEBUG nova.network.neutron [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 663.125862] env[63021]: ERROR nova.compute.manager [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 31850b77-7688-482d-91fb-d74fdd83e293, please check neutron logs for more information. [ 663.125862] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 663.125862] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 663.125862] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 663.125862] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 663.125862] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 663.125862] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 663.125862] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 663.125862] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.125862] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 663.125862] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.125862] env[63021]: ERROR nova.compute.manager raise self.value [ 663.125862] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 663.125862] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 663.125862] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.125862] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 663.126306] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.126306] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 663.126306] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 31850b77-7688-482d-91fb-d74fdd83e293, please check neutron logs for more information. [ 663.126306] env[63021]: ERROR nova.compute.manager [ 663.126306] env[63021]: Traceback (most recent call last): [ 663.126306] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 663.126306] env[63021]: listener.cb(fileno) [ 663.126306] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 663.126306] env[63021]: result = function(*args, **kwargs) [ 663.126306] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 663.126306] env[63021]: return func(*args, **kwargs) [ 663.126306] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 663.126306] env[63021]: raise e [ 663.126306] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 663.126306] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 663.126306] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 663.126306] env[63021]: created_port_ids = self._update_ports_for_instance( [ 663.126306] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 663.126306] env[63021]: with excutils.save_and_reraise_exception(): [ 663.126306] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.126306] env[63021]: self.force_reraise() [ 663.126306] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.126306] env[63021]: raise self.value [ 663.126306] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 663.126306] env[63021]: updated_port = self._update_port( [ 663.126306] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.126306] env[63021]: _ensure_no_port_binding_failure(port) [ 663.126306] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.126306] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 663.126995] env[63021]: nova.exception.PortBindingFailed: Binding failed for port 31850b77-7688-482d-91fb-d74fdd83e293, please check neutron logs for more information. [ 663.126995] env[63021]: Removing descriptor: 17 [ 663.126995] env[63021]: ERROR nova.compute.manager [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 31850b77-7688-482d-91fb-d74fdd83e293, please check neutron logs for more information. [ 663.126995] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Traceback (most recent call last): [ 663.126995] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 663.126995] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] yield resources [ 663.126995] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 663.126995] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] self.driver.spawn(context, instance, image_meta, [ 663.126995] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 663.126995] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 663.126995] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 663.126995] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] vm_ref = self.build_virtual_machine(instance, [ 663.127301] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 663.127301] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] vif_infos = vmwarevif.get_vif_info(self._session, [ 663.127301] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 663.127301] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] for vif in network_info: [ 663.127301] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 663.127301] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] return self._sync_wrapper(fn, *args, **kwargs) [ 663.127301] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 663.127301] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] self.wait() [ 663.127301] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 663.127301] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] self[:] = self._gt.wait() [ 663.127301] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 663.127301] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] return self._exit_event.wait() [ 663.127301] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 663.127602] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] result = hub.switch() [ 663.127602] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 663.127602] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] return self.greenlet.switch() [ 663.127602] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 663.127602] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] result = function(*args, **kwargs) [ 663.127602] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 663.127602] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] return func(*args, **kwargs) [ 663.127602] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 663.127602] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] raise e [ 663.127602] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 663.127602] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] nwinfo = self.network_api.allocate_for_instance( [ 663.127602] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 663.127602] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] created_port_ids = self._update_ports_for_instance( [ 663.128019] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 663.128019] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] with excutils.save_and_reraise_exception(): [ 663.128019] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.128019] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] self.force_reraise() [ 663.128019] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.128019] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] raise self.value [ 663.128019] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 663.128019] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] updated_port = self._update_port( [ 663.128019] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.128019] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] _ensure_no_port_binding_failure(port) [ 663.128019] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.128019] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] raise exception.PortBindingFailed(port_id=port['id']) [ 663.128311] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] nova.exception.PortBindingFailed: Binding failed for port 31850b77-7688-482d-91fb-d74fdd83e293, please check neutron logs for more information. [ 663.128311] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] [ 663.128311] env[63021]: INFO nova.compute.manager [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Terminating instance [ 663.130507] env[63021]: DEBUG oslo_concurrency.lockutils [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Acquiring lock "refresh_cache-05caabab-8c7d-4d27-8eb3-e7e1c2947aa6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.200760] env[63021]: DEBUG nova.policy [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6d9d1625692d4081855d416c91d9cb87', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b1c3fa55c2344a83a0f33ce8f6a6098d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 663.600389] env[63021]: DEBUG nova.network.neutron [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Successfully created port: da5bf4a2-c61a-4810-8e49-677fa4a48055 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 663.626519] env[63021]: DEBUG nova.compute.manager [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 663.665094] env[63021]: DEBUG nova.network.neutron [req-3479c9eb-10e7-4710-9de5-ead7a8b8b91a req-beb926b8-732d-45d3-8831-6ba74e5d1a33 service nova] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 663.745995] env[63021]: DEBUG nova.network.neutron [req-3479c9eb-10e7-4710-9de5-ead7a8b8b91a req-beb926b8-732d-45d3-8831-6ba74e5d1a33 service nova] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.067193] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8814f9db-0020-48a1-9076-85408fcd27f9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.075677] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b73d124-9da0-4882-8290-aa8dec4aa00a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.106461] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2677f22a-ce7f-4be2-9b37-11cd1676c143 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.114212] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b656a22a-ec22-47c6-aa2d-ca4d03eb5e5f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.127946] env[63021]: DEBUG nova.compute.provider_tree [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 664.253964] env[63021]: DEBUG oslo_concurrency.lockutils [req-3479c9eb-10e7-4710-9de5-ead7a8b8b91a req-beb926b8-732d-45d3-8831-6ba74e5d1a33 service nova] Releasing lock "refresh_cache-05caabab-8c7d-4d27-8eb3-e7e1c2947aa6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.254510] env[63021]: DEBUG oslo_concurrency.lockutils [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Acquired lock "refresh_cache-05caabab-8c7d-4d27-8eb3-e7e1c2947aa6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.254753] env[63021]: DEBUG nova.network.neutron [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 664.601943] env[63021]: DEBUG nova.compute.manager [req-e3009cb1-c9ad-4e43-9aff-4e7f3509fe5a req-596f4f6b-7516-41ea-bfa1-7111c899f627 service nova] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Received event network-changed-da5bf4a2-c61a-4810-8e49-677fa4a48055 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 664.602240] env[63021]: DEBUG nova.compute.manager [req-e3009cb1-c9ad-4e43-9aff-4e7f3509fe5a req-596f4f6b-7516-41ea-bfa1-7111c899f627 service nova] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Refreshing instance network info cache due to event network-changed-da5bf4a2-c61a-4810-8e49-677fa4a48055. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 664.602468] env[63021]: DEBUG oslo_concurrency.lockutils [req-e3009cb1-c9ad-4e43-9aff-4e7f3509fe5a req-596f4f6b-7516-41ea-bfa1-7111c899f627 service nova] Acquiring lock "refresh_cache-39f51f90-a5a5-4479-881c-b5b23a3bf01a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.602771] env[63021]: DEBUG oslo_concurrency.lockutils [req-e3009cb1-c9ad-4e43-9aff-4e7f3509fe5a req-596f4f6b-7516-41ea-bfa1-7111c899f627 service nova] Acquired lock "refresh_cache-39f51f90-a5a5-4479-881c-b5b23a3bf01a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.602880] env[63021]: DEBUG nova.network.neutron [req-e3009cb1-c9ad-4e43-9aff-4e7f3509fe5a req-596f4f6b-7516-41ea-bfa1-7111c899f627 service nova] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Refreshing network info cache for port da5bf4a2-c61a-4810-8e49-677fa4a48055 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 664.630268] env[63021]: DEBUG nova.scheduler.client.report [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 664.637088] env[63021]: DEBUG nova.compute.manager [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 664.665023] env[63021]: DEBUG nova.virt.hardware [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 664.665023] env[63021]: DEBUG nova.virt.hardware [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 664.665023] env[63021]: DEBUG nova.virt.hardware [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 664.665675] env[63021]: DEBUG nova.virt.hardware [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 664.665675] env[63021]: DEBUG nova.virt.hardware [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 664.665675] env[63021]: DEBUG nova.virt.hardware [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 664.665675] env[63021]: DEBUG nova.virt.hardware [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 664.665675] env[63021]: DEBUG nova.virt.hardware [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 664.665997] env[63021]: DEBUG nova.virt.hardware [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 664.665997] env[63021]: DEBUG nova.virt.hardware [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 664.665997] env[63021]: DEBUG nova.virt.hardware [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 664.666808] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afd349db-514a-43c0-96a2-c065d61ea206 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.677563] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a817be55-5614-485c-9a67-7c615d77d321 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.695515] env[63021]: ERROR nova.compute.manager [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port da5bf4a2-c61a-4810-8e49-677fa4a48055, please check neutron logs for more information. [ 664.695515] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 664.695515] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 664.695515] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 664.695515] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 664.695515] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 664.695515] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 664.695515] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 664.695515] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.695515] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 664.695515] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.695515] env[63021]: ERROR nova.compute.manager raise self.value [ 664.695515] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 664.695515] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 664.695515] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.695515] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 664.695996] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.695996] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 664.695996] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port da5bf4a2-c61a-4810-8e49-677fa4a48055, please check neutron logs for more information. [ 664.695996] env[63021]: ERROR nova.compute.manager [ 664.695996] env[63021]: Traceback (most recent call last): [ 664.695996] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 664.695996] env[63021]: listener.cb(fileno) [ 664.695996] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 664.695996] env[63021]: result = function(*args, **kwargs) [ 664.695996] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 664.695996] env[63021]: return func(*args, **kwargs) [ 664.695996] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 664.695996] env[63021]: raise e [ 664.695996] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 664.695996] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 664.695996] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 664.695996] env[63021]: created_port_ids = self._update_ports_for_instance( [ 664.695996] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 664.695996] env[63021]: with excutils.save_and_reraise_exception(): [ 664.695996] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.695996] env[63021]: self.force_reraise() [ 664.695996] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.695996] env[63021]: raise self.value [ 664.695996] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 664.695996] env[63021]: updated_port = self._update_port( [ 664.695996] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.695996] env[63021]: _ensure_no_port_binding_failure(port) [ 664.695996] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.695996] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 664.696957] env[63021]: nova.exception.PortBindingFailed: Binding failed for port da5bf4a2-c61a-4810-8e49-677fa4a48055, please check neutron logs for more information. [ 664.696957] env[63021]: Removing descriptor: 18 [ 664.696957] env[63021]: ERROR nova.compute.manager [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port da5bf4a2-c61a-4810-8e49-677fa4a48055, please check neutron logs for more information. [ 664.696957] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Traceback (most recent call last): [ 664.696957] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 664.696957] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] yield resources [ 664.696957] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 664.696957] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] self.driver.spawn(context, instance, image_meta, [ 664.696957] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 664.696957] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 664.696957] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 664.696957] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] vm_ref = self.build_virtual_machine(instance, [ 664.697419] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 664.697419] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] vif_infos = vmwarevif.get_vif_info(self._session, [ 664.697419] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 664.697419] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] for vif in network_info: [ 664.697419] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 664.697419] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] return self._sync_wrapper(fn, *args, **kwargs) [ 664.697419] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 664.697419] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] self.wait() [ 664.697419] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 664.697419] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] self[:] = self._gt.wait() [ 664.697419] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 664.697419] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] return self._exit_event.wait() [ 664.697419] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 664.697778] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] result = hub.switch() [ 664.697778] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 664.697778] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] return self.greenlet.switch() [ 664.697778] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 664.697778] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] result = function(*args, **kwargs) [ 664.697778] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 664.697778] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] return func(*args, **kwargs) [ 664.697778] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 664.697778] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] raise e [ 664.697778] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 664.697778] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] nwinfo = self.network_api.allocate_for_instance( [ 664.697778] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 664.697778] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] created_port_ids = self._update_ports_for_instance( [ 664.698267] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 664.698267] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] with excutils.save_and_reraise_exception(): [ 664.698267] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.698267] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] self.force_reraise() [ 664.698267] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.698267] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] raise self.value [ 664.698267] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 664.698267] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] updated_port = self._update_port( [ 664.698267] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.698267] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] _ensure_no_port_binding_failure(port) [ 664.698267] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.698267] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] raise exception.PortBindingFailed(port_id=port['id']) [ 664.698625] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] nova.exception.PortBindingFailed: Binding failed for port da5bf4a2-c61a-4810-8e49-677fa4a48055, please check neutron logs for more information. [ 664.698625] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] [ 664.698625] env[63021]: INFO nova.compute.manager [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Terminating instance [ 664.698855] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Acquiring lock "refresh_cache-39f51f90-a5a5-4479-881c-b5b23a3bf01a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.780284] env[63021]: DEBUG nova.network.neutron [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.861444] env[63021]: DEBUG nova.network.neutron [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.126099] env[63021]: DEBUG nova.network.neutron [req-e3009cb1-c9ad-4e43-9aff-4e7f3509fe5a req-596f4f6b-7516-41ea-bfa1-7111c899f627 service nova] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 665.135235] env[63021]: DEBUG oslo_concurrency.lockutils [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.523s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.135746] env[63021]: DEBUG nova.compute.manager [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 665.138331] env[63021]: DEBUG oslo_concurrency.lockutils [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.376s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.139765] env[63021]: INFO nova.compute.claims [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 665.145699] env[63021]: DEBUG nova.compute.manager [req-04230a08-bce4-4e40-a8a5-44f1399e9ea3 req-8f9058a9-692f-47f3-834a-79d263e71278 service nova] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Received event network-vif-deleted-31850b77-7688-482d-91fb-d74fdd83e293 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 665.296820] env[63021]: DEBUG nova.network.neutron [req-e3009cb1-c9ad-4e43-9aff-4e7f3509fe5a req-596f4f6b-7516-41ea-bfa1-7111c899f627 service nova] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.364345] env[63021]: DEBUG oslo_concurrency.lockutils [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Releasing lock "refresh_cache-05caabab-8c7d-4d27-8eb3-e7e1c2947aa6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 665.364769] env[63021]: DEBUG nova.compute.manager [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 665.364991] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 665.366027] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8d908735-8252-4bbd-bf05-244e870f7d42 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.374941] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b127cba5-1d12-4be2-8bed-0a861cd37d04 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.399737] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6 could not be found. [ 665.399978] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 665.400177] env[63021]: INFO nova.compute.manager [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 665.400466] env[63021]: DEBUG oslo.service.loopingcall [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 665.400683] env[63021]: DEBUG nova.compute.manager [-] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 665.400781] env[63021]: DEBUG nova.network.neutron [-] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 665.422198] env[63021]: DEBUG nova.network.neutron [-] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 665.644306] env[63021]: DEBUG nova.compute.utils [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 665.647636] env[63021]: DEBUG nova.compute.manager [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Not allocating networking since 'none' was specified. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 665.799432] env[63021]: DEBUG oslo_concurrency.lockutils [req-e3009cb1-c9ad-4e43-9aff-4e7f3509fe5a req-596f4f6b-7516-41ea-bfa1-7111c899f627 service nova] Releasing lock "refresh_cache-39f51f90-a5a5-4479-881c-b5b23a3bf01a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 665.799915] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Acquired lock "refresh_cache-39f51f90-a5a5-4479-881c-b5b23a3bf01a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.800163] env[63021]: DEBUG nova.network.neutron [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 665.926390] env[63021]: DEBUG nova.network.neutron [-] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.148305] env[63021]: DEBUG nova.compute.manager [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 666.320487] env[63021]: DEBUG nova.network.neutron [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 666.431825] env[63021]: INFO nova.compute.manager [-] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Took 1.03 seconds to deallocate network for instance. [ 666.434071] env[63021]: DEBUG nova.compute.claims [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 666.434307] env[63021]: DEBUG oslo_concurrency.lockutils [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 666.451737] env[63021]: DEBUG nova.network.neutron [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.558955] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90d92be4-13a6-47fb-bf8b-5cc7202f6303 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.571033] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aa29d91-65ca-4162-96f0-1dbb53d6880d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.605047] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d23c0e9-3cd4-45f1-91eb-16d2af6f45f4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.613125] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be461132-584c-4a50-8f1f-f9724df2b047 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.627900] env[63021]: DEBUG nova.compute.provider_tree [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 666.712383] env[63021]: DEBUG nova.compute.manager [req-b817a6b8-50f8-4a9d-856b-852ae91911f7 req-5ae1b690-e00a-48cb-aff0-8812a8de6ec0 service nova] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Received event network-vif-deleted-da5bf4a2-c61a-4810-8e49-677fa4a48055 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 666.954142] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Releasing lock "refresh_cache-39f51f90-a5a5-4479-881c-b5b23a3bf01a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 666.954572] env[63021]: DEBUG nova.compute.manager [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 666.954760] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 666.955070] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3d483268-0db6-46f1-963a-5002ab21c731 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.965054] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f1b4078-3a26-41a4-94b5-b8ae806233fe {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.987363] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 39f51f90-a5a5-4479-881c-b5b23a3bf01a could not be found. [ 666.987600] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 666.987783] env[63021]: INFO nova.compute.manager [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Took 0.03 seconds to destroy the instance on the hypervisor. [ 666.988061] env[63021]: DEBUG oslo.service.loopingcall [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 666.988306] env[63021]: DEBUG nova.compute.manager [-] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 666.988362] env[63021]: DEBUG nova.network.neutron [-] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 667.003407] env[63021]: DEBUG nova.network.neutron [-] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 667.133299] env[63021]: DEBUG nova.scheduler.client.report [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 667.164489] env[63021]: DEBUG nova.compute.manager [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 667.196429] env[63021]: DEBUG nova.virt.hardware [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 667.196429] env[63021]: DEBUG nova.virt.hardware [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 667.196598] env[63021]: DEBUG nova.virt.hardware [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 667.196670] env[63021]: DEBUG nova.virt.hardware [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 667.196814] env[63021]: DEBUG nova.virt.hardware [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 667.196958] env[63021]: DEBUG nova.virt.hardware [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 667.197176] env[63021]: DEBUG nova.virt.hardware [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 667.197332] env[63021]: DEBUG nova.virt.hardware [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 667.197494] env[63021]: DEBUG nova.virt.hardware [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 667.197652] env[63021]: DEBUG nova.virt.hardware [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 667.197822] env[63021]: DEBUG nova.virt.hardware [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 667.198738] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e449a9c4-8e3a-4c62-9deb-8158eac44dbb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.210523] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ae377f7-a90f-40da-9f87-5117e863213e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.228890] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Instance VIF info [] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 667.234551] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Creating folder: Project (d8c0d87fb9f746c8a65061207bd7837b). Parent ref: group-v277447. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 667.235272] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-23ac7694-76f7-49ca-a365-ffb5b9f2f94a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.245973] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Created folder: Project (d8c0d87fb9f746c8a65061207bd7837b) in parent group-v277447. [ 667.245973] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Creating folder: Instances. Parent ref: group-v277460. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 667.245973] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-653a6dee-da5b-4fa9-9460-ba809e094d73 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.253981] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Created folder: Instances in parent group-v277460. [ 667.254246] env[63021]: DEBUG oslo.service.loopingcall [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 667.254434] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 667.254636] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2472ad60-e914-43e1-baed-57353fa646f5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.270986] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 667.270986] env[63021]: value = "task-1293581" [ 667.270986] env[63021]: _type = "Task" [ 667.270986] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.279491] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293581, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.506547] env[63021]: DEBUG nova.network.neutron [-] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.640323] env[63021]: DEBUG oslo_concurrency.lockutils [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.502s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.640861] env[63021]: DEBUG nova.compute.manager [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 667.643469] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.748s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 667.783089] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293581, 'name': CreateVM_Task, 'duration_secs': 0.247194} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.783282] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 667.783701] env[63021]: DEBUG oslo_concurrency.lockutils [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.783858] env[63021]: DEBUG oslo_concurrency.lockutils [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.784187] env[63021]: DEBUG oslo_concurrency.lockutils [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 667.784437] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2580171f-0e4a-475b-9924-e927189c249b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.789131] env[63021]: DEBUG oslo_vmware.api [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Waiting for the task: (returnval){ [ 667.789131] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52f17747-7d20-7727-8ba0-5209bf909504" [ 667.789131] env[63021]: _type = "Task" [ 667.789131] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.799187] env[63021]: DEBUG oslo_vmware.api [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52f17747-7d20-7727-8ba0-5209bf909504, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.008904] env[63021]: INFO nova.compute.manager [-] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Took 1.02 seconds to deallocate network for instance. [ 668.011541] env[63021]: DEBUG nova.compute.claims [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 668.011771] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.148450] env[63021]: DEBUG nova.compute.utils [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 668.154668] env[63021]: DEBUG nova.compute.manager [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 668.154858] env[63021]: DEBUG nova.network.neutron [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 668.202434] env[63021]: DEBUG nova.policy [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '16118a00939847db9f966f2942d3c2e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c2698b5193df491a93a62e91202b3fb3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 668.307027] env[63021]: DEBUG oslo_vmware.api [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52f17747-7d20-7727-8ba0-5209bf909504, 'name': SearchDatastore_Task, 'duration_secs': 0.010637} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.307027] env[63021]: DEBUG oslo_concurrency.lockutils [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 668.307027] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 668.307027] env[63021]: DEBUG oslo_concurrency.lockutils [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 668.307408] env[63021]: DEBUG oslo_concurrency.lockutils [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.307408] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 668.307408] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-861f4194-3d64-448c-8e5b-047ca5982e1b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.316895] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 668.317684] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 668.317846] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4793f19b-e6d1-45e4-a333-8cc0ba64bbf4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.323101] env[63021]: DEBUG oslo_vmware.api [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Waiting for the task: (returnval){ [ 668.323101] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]527de3ae-b1e3-0bd5-6d09-e9873eeab060" [ 668.323101] env[63021]: _type = "Task" [ 668.323101] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.336717] env[63021]: DEBUG oslo_vmware.api [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]527de3ae-b1e3-0bd5-6d09-e9873eeab060, 'name': SearchDatastore_Task, 'duration_secs': 0.008401} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.337593] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af88150c-5749-4085-a2dc-375eadec1ffb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.343727] env[63021]: DEBUG oslo_vmware.api [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Waiting for the task: (returnval){ [ 668.343727] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]523bb4de-004f-4cf3-e25c-ba995ab3ad16" [ 668.343727] env[63021]: _type = "Task" [ 668.343727] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.351078] env[63021]: DEBUG oslo_vmware.api [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]523bb4de-004f-4cf3-e25c-ba995ab3ad16, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.503976] env[63021]: DEBUG nova.network.neutron [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Successfully created port: 559cf9af-5956-4bb5-8fec-4f748c72d59b {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 668.531693] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77a32ae8-625c-486f-a9df-d43ab8f0e2e8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.538909] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d833a55-f184-4f6a-b737-8088f2b94f52 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.570953] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d6360d0-c8c0-4946-a925-231bb646d9bc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.578025] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Acquiring lock "9f3b0e67-24fc-4625-81f4-050055906c19" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.578252] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Lock "9f3b0e67-24fc-4625-81f4-050055906c19" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.582684] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccf249de-48cd-411f-b00d-262a0532e324 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.599366] env[63021]: DEBUG nova.compute.provider_tree [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 668.655055] env[63021]: DEBUG nova.compute.manager [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 668.861135] env[63021]: DEBUG oslo_vmware.api [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]523bb4de-004f-4cf3-e25c-ba995ab3ad16, 'name': SearchDatastore_Task, 'duration_secs': 0.008921} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.861135] env[63021]: DEBUG oslo_concurrency.lockutils [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 668.861135] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 77fd1199-33b5-4040-a0ba-dd881da1224c/77fd1199-33b5-4040-a0ba-dd881da1224c.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 668.861135] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-54d333aa-4cdc-4249-8fef-b1e7f7f0110f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.866657] env[63021]: DEBUG oslo_vmware.api [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Waiting for the task: (returnval){ [ 668.866657] env[63021]: value = "task-1293582" [ 668.866657] env[63021]: _type = "Task" [ 668.866657] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.875327] env[63021]: DEBUG oslo_vmware.api [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Task: {'id': task-1293582, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.103308] env[63021]: DEBUG nova.scheduler.client.report [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 669.247156] env[63021]: DEBUG nova.compute.manager [req-0ac04f48-1572-4328-8aa4-76a3bf6fd395 req-f26a31de-a142-4918-96e4-c8756e7b9403 service nova] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Received event network-changed-559cf9af-5956-4bb5-8fec-4f748c72d59b {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 669.247156] env[63021]: DEBUG nova.compute.manager [req-0ac04f48-1572-4328-8aa4-76a3bf6fd395 req-f26a31de-a142-4918-96e4-c8756e7b9403 service nova] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Refreshing instance network info cache due to event network-changed-559cf9af-5956-4bb5-8fec-4f748c72d59b. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 669.247156] env[63021]: DEBUG oslo_concurrency.lockutils [req-0ac04f48-1572-4328-8aa4-76a3bf6fd395 req-f26a31de-a142-4918-96e4-c8756e7b9403 service nova] Acquiring lock "refresh_cache-4e6dcd18-4814-4cf1-966b-7d853858e650" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.247156] env[63021]: DEBUG oslo_concurrency.lockutils [req-0ac04f48-1572-4328-8aa4-76a3bf6fd395 req-f26a31de-a142-4918-96e4-c8756e7b9403 service nova] Acquired lock "refresh_cache-4e6dcd18-4814-4cf1-966b-7d853858e650" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.247156] env[63021]: DEBUG nova.network.neutron [req-0ac04f48-1572-4328-8aa4-76a3bf6fd395 req-f26a31de-a142-4918-96e4-c8756e7b9403 service nova] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Refreshing network info cache for port 559cf9af-5956-4bb5-8fec-4f748c72d59b {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 669.377808] env[63021]: DEBUG oslo_vmware.api [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Task: {'id': task-1293582, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.378790] env[63021]: ERROR nova.compute.manager [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 559cf9af-5956-4bb5-8fec-4f748c72d59b, please check neutron logs for more information. [ 669.378790] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 669.378790] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.378790] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 669.378790] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 669.378790] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 669.378790] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 669.378790] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 669.378790] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.378790] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 669.378790] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.378790] env[63021]: ERROR nova.compute.manager raise self.value [ 669.378790] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 669.378790] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 669.378790] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.378790] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 669.379311] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.379311] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 669.379311] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 559cf9af-5956-4bb5-8fec-4f748c72d59b, please check neutron logs for more information. [ 669.379311] env[63021]: ERROR nova.compute.manager [ 669.379311] env[63021]: Traceback (most recent call last): [ 669.379311] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 669.379311] env[63021]: listener.cb(fileno) [ 669.379311] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.379311] env[63021]: result = function(*args, **kwargs) [ 669.379311] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 669.379311] env[63021]: return func(*args, **kwargs) [ 669.379311] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 669.379311] env[63021]: raise e [ 669.379311] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.379311] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 669.379311] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 669.379311] env[63021]: created_port_ids = self._update_ports_for_instance( [ 669.379311] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 669.379311] env[63021]: with excutils.save_and_reraise_exception(): [ 669.379311] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.379311] env[63021]: self.force_reraise() [ 669.379311] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.379311] env[63021]: raise self.value [ 669.379311] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 669.379311] env[63021]: updated_port = self._update_port( [ 669.379311] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.379311] env[63021]: _ensure_no_port_binding_failure(port) [ 669.379311] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.379311] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 669.380407] env[63021]: nova.exception.PortBindingFailed: Binding failed for port 559cf9af-5956-4bb5-8fec-4f748c72d59b, please check neutron logs for more information. [ 669.380407] env[63021]: Removing descriptor: 18 [ 669.608884] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.965s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.609606] env[63021]: ERROR nova.compute.manager [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 97e064ac-be35-4a14-960e-34af1585617f, please check neutron logs for more information. [ 669.609606] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Traceback (most recent call last): [ 669.609606] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 669.609606] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] self.driver.spawn(context, instance, image_meta, [ 669.609606] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 669.609606] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 669.609606] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 669.609606] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] vm_ref = self.build_virtual_machine(instance, [ 669.609606] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 669.609606] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] vif_infos = vmwarevif.get_vif_info(self._session, [ 669.609606] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 669.609973] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] for vif in network_info: [ 669.609973] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 669.609973] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] return self._sync_wrapper(fn, *args, **kwargs) [ 669.609973] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 669.609973] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] self.wait() [ 669.609973] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 669.609973] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] self[:] = self._gt.wait() [ 669.609973] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 669.609973] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] return self._exit_event.wait() [ 669.609973] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 669.609973] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] result = hub.switch() [ 669.609973] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 669.609973] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] return self.greenlet.switch() [ 669.610348] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.610348] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] result = function(*args, **kwargs) [ 669.610348] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 669.610348] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] return func(*args, **kwargs) [ 669.610348] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 669.610348] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] raise e [ 669.610348] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.610348] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] nwinfo = self.network_api.allocate_for_instance( [ 669.610348] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 669.610348] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] created_port_ids = self._update_ports_for_instance( [ 669.610348] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 669.610348] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] with excutils.save_and_reraise_exception(): [ 669.610348] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.610712] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] self.force_reraise() [ 669.610712] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.610712] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] raise self.value [ 669.610712] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 669.610712] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] updated_port = self._update_port( [ 669.610712] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.610712] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] _ensure_no_port_binding_failure(port) [ 669.610712] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.610712] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] raise exception.PortBindingFailed(port_id=port['id']) [ 669.610712] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] nova.exception.PortBindingFailed: Binding failed for port 97e064ac-be35-4a14-960e-34af1585617f, please check neutron logs for more information. [ 669.610712] env[63021]: ERROR nova.compute.manager [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] [ 669.611038] env[63021]: DEBUG nova.compute.utils [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Binding failed for port 97e064ac-be35-4a14-960e-34af1585617f, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 669.611774] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.360s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.615741] env[63021]: DEBUG nova.compute.manager [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Build of instance 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c was re-scheduled: Binding failed for port 97e064ac-be35-4a14-960e-34af1585617f, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 669.615867] env[63021]: DEBUG nova.compute.manager [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 669.617032] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "refresh_cache-12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.617032] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquired lock "refresh_cache-12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.617032] env[63021]: DEBUG nova.network.neutron [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 669.665054] env[63021]: DEBUG nova.compute.manager [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 669.689454] env[63021]: DEBUG nova.virt.hardware [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 669.689745] env[63021]: DEBUG nova.virt.hardware [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 669.689907] env[63021]: DEBUG nova.virt.hardware [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 669.690115] env[63021]: DEBUG nova.virt.hardware [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 669.690266] env[63021]: DEBUG nova.virt.hardware [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 669.690409] env[63021]: DEBUG nova.virt.hardware [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 669.690608] env[63021]: DEBUG nova.virt.hardware [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 669.690765] env[63021]: DEBUG nova.virt.hardware [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 669.690925] env[63021]: DEBUG nova.virt.hardware [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 669.691094] env[63021]: DEBUG nova.virt.hardware [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 669.691265] env[63021]: DEBUG nova.virt.hardware [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 669.692652] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed16bdd0-fe13-4d04-8726-0ee6e697bd81 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.700444] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab8fb997-a54d-4de6-9a5f-8cf959442052 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.714246] env[63021]: ERROR nova.compute.manager [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 559cf9af-5956-4bb5-8fec-4f748c72d59b, please check neutron logs for more information. [ 669.714246] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Traceback (most recent call last): [ 669.714246] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 669.714246] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] yield resources [ 669.714246] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 669.714246] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] self.driver.spawn(context, instance, image_meta, [ 669.714246] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 669.714246] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] self._vmops.spawn(context, instance, image_meta, injected_files, [ 669.714246] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 669.714246] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] vm_ref = self.build_virtual_machine(instance, [ 669.714246] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 669.714648] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] vif_infos = vmwarevif.get_vif_info(self._session, [ 669.714648] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 669.714648] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] for vif in network_info: [ 669.714648] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 669.714648] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] return self._sync_wrapper(fn, *args, **kwargs) [ 669.714648] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 669.714648] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] self.wait() [ 669.714648] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 669.714648] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] self[:] = self._gt.wait() [ 669.714648] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 669.714648] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] return self._exit_event.wait() [ 669.714648] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 669.714648] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] current.throw(*self._exc) [ 669.715080] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.715080] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] result = function(*args, **kwargs) [ 669.715080] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 669.715080] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] return func(*args, **kwargs) [ 669.715080] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 669.715080] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] raise e [ 669.715080] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.715080] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] nwinfo = self.network_api.allocate_for_instance( [ 669.715080] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 669.715080] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] created_port_ids = self._update_ports_for_instance( [ 669.715080] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 669.715080] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] with excutils.save_and_reraise_exception(): [ 669.715080] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.715448] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] self.force_reraise() [ 669.715448] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.715448] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] raise self.value [ 669.715448] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 669.715448] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] updated_port = self._update_port( [ 669.715448] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.715448] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] _ensure_no_port_binding_failure(port) [ 669.715448] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.715448] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] raise exception.PortBindingFailed(port_id=port['id']) [ 669.715448] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] nova.exception.PortBindingFailed: Binding failed for port 559cf9af-5956-4bb5-8fec-4f748c72d59b, please check neutron logs for more information. [ 669.715448] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] [ 669.715448] env[63021]: INFO nova.compute.manager [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Terminating instance [ 669.716542] env[63021]: DEBUG oslo_concurrency.lockutils [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Acquiring lock "refresh_cache-4e6dcd18-4814-4cf1-966b-7d853858e650" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.763654] env[63021]: DEBUG nova.network.neutron [req-0ac04f48-1572-4328-8aa4-76a3bf6fd395 req-f26a31de-a142-4918-96e4-c8756e7b9403 service nova] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 669.839386] env[63021]: DEBUG nova.network.neutron [req-0ac04f48-1572-4328-8aa4-76a3bf6fd395 req-f26a31de-a142-4918-96e4-c8756e7b9403 service nova] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.878543] env[63021]: DEBUG oslo_vmware.api [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Task: {'id': task-1293582, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.589894} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.878804] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 77fd1199-33b5-4040-a0ba-dd881da1224c/77fd1199-33b5-4040-a0ba-dd881da1224c.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 669.879022] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 669.879265] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a18c6d98-dfc7-4ab7-8ebe-7d735ab90785 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.885757] env[63021]: DEBUG oslo_vmware.api [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Waiting for the task: (returnval){ [ 669.885757] env[63021]: value = "task-1293583" [ 669.885757] env[63021]: _type = "Task" [ 669.885757] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.893545] env[63021]: DEBUG oslo_vmware.api [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Task: {'id': task-1293583, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.137824] env[63021]: DEBUG nova.network.neutron [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 670.213852] env[63021]: DEBUG nova.network.neutron [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.342974] env[63021]: DEBUG oslo_concurrency.lockutils [req-0ac04f48-1572-4328-8aa4-76a3bf6fd395 req-f26a31de-a142-4918-96e4-c8756e7b9403 service nova] Releasing lock "refresh_cache-4e6dcd18-4814-4cf1-966b-7d853858e650" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.343415] env[63021]: DEBUG oslo_concurrency.lockutils [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Acquired lock "refresh_cache-4e6dcd18-4814-4cf1-966b-7d853858e650" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.343600] env[63021]: DEBUG nova.network.neutron [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 670.400381] env[63021]: DEBUG oslo_vmware.api [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Task: {'id': task-1293583, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069716} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.400381] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 670.400381] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfe6df75-91d2-4722-a2a2-0bbc5fae13ee {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.421190] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Reconfiguring VM instance instance-0000001e to attach disk [datastore1] 77fd1199-33b5-4040-a0ba-dd881da1224c/77fd1199-33b5-4040-a0ba-dd881da1224c.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 670.422108] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-98152a96-b45e-46cf-8d2b-a48cdde51fbd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.441192] env[63021]: DEBUG oslo_vmware.api [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Waiting for the task: (returnval){ [ 670.441192] env[63021]: value = "task-1293584" [ 670.441192] env[63021]: _type = "Task" [ 670.441192] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.450192] env[63021]: DEBUG oslo_vmware.api [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Task: {'id': task-1293584, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.475352] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62561870-4a6c-4ff6-8a6a-ff654909ef86 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.482273] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77496866-a9f5-45bc-ba33-e59139529d4e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.512282] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df9f4f72-80a1-4d8e-a1f7-27c122d889b3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.519654] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f106f0d0-46c1-4d06-8313-398ecfbac96e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.535586] env[63021]: DEBUG nova.compute.provider_tree [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 670.718396] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Releasing lock "refresh_cache-12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.718709] env[63021]: DEBUG nova.compute.manager [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 670.718915] env[63021]: DEBUG nova.compute.manager [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 670.719504] env[63021]: DEBUG nova.network.neutron [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 670.734793] env[63021]: DEBUG nova.network.neutron [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 670.861524] env[63021]: DEBUG nova.network.neutron [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 670.934290] env[63021]: DEBUG nova.network.neutron [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.951421] env[63021]: DEBUG oslo_vmware.api [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Task: {'id': task-1293584, 'name': ReconfigVM_Task, 'duration_secs': 0.281577} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.951766] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Reconfigured VM instance instance-0000001e to attach disk [datastore1] 77fd1199-33b5-4040-a0ba-dd881da1224c/77fd1199-33b5-4040-a0ba-dd881da1224c.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 670.952430] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8e66fdc9-7733-4481-baa6-b5e1425ce5c6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.958834] env[63021]: DEBUG oslo_vmware.api [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Waiting for the task: (returnval){ [ 670.958834] env[63021]: value = "task-1293585" [ 670.958834] env[63021]: _type = "Task" [ 670.958834] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.966875] env[63021]: DEBUG oslo_vmware.api [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Task: {'id': task-1293585, 'name': Rename_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.037055] env[63021]: DEBUG nova.scheduler.client.report [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 671.237183] env[63021]: DEBUG nova.network.neutron [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.270377] env[63021]: DEBUG nova.compute.manager [req-cd6d5421-81d3-4175-a8c4-b3ab099d97ed req-17fe4756-22ba-40bc-96b0-781bdc34cad5 service nova] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Received event network-vif-deleted-559cf9af-5956-4bb5-8fec-4f748c72d59b {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 671.439024] env[63021]: DEBUG oslo_concurrency.lockutils [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Releasing lock "refresh_cache-4e6dcd18-4814-4cf1-966b-7d853858e650" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.439024] env[63021]: DEBUG nova.compute.manager [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 671.439024] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 671.439024] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ea9dd4ff-0962-460e-a018-8378315caa1e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.447088] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee386242-b9ea-4aca-bf0a-a264591fc425 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.467340] env[63021]: DEBUG oslo_vmware.api [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Task: {'id': task-1293585, 'name': Rename_Task, 'duration_secs': 0.13575} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.471361] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 671.472062] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4e6dcd18-4814-4cf1-966b-7d853858e650 could not be found. [ 671.472354] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 671.472625] env[63021]: INFO nova.compute.manager [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Took 0.04 seconds to destroy the instance on the hypervisor. [ 671.472984] env[63021]: DEBUG oslo.service.loopingcall [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 671.473299] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-737a10dd-5f64-4e46-a7d4-0f357eb60d59 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.474894] env[63021]: DEBUG nova.compute.manager [-] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 671.475109] env[63021]: DEBUG nova.network.neutron [-] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 671.481290] env[63021]: DEBUG oslo_vmware.api [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Waiting for the task: (returnval){ [ 671.481290] env[63021]: value = "task-1293586" [ 671.481290] env[63021]: _type = "Task" [ 671.481290] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.489497] env[63021]: DEBUG oslo_vmware.api [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Task: {'id': task-1293586, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.490247] env[63021]: DEBUG nova.network.neutron [-] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 671.542053] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.930s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 671.542745] env[63021]: ERROR nova.compute.manager [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0b0dcc76-2a8c-4f06-b32e-22e39f072235, please check neutron logs for more information. [ 671.542745] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Traceback (most recent call last): [ 671.542745] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 671.542745] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] self.driver.spawn(context, instance, image_meta, [ 671.542745] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 671.542745] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] self._vmops.spawn(context, instance, image_meta, injected_files, [ 671.542745] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 671.542745] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] vm_ref = self.build_virtual_machine(instance, [ 671.542745] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 671.542745] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] vif_infos = vmwarevif.get_vif_info(self._session, [ 671.542745] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 671.543168] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] for vif in network_info: [ 671.543168] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 671.543168] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] return self._sync_wrapper(fn, *args, **kwargs) [ 671.543168] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 671.543168] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] self.wait() [ 671.543168] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 671.543168] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] self[:] = self._gt.wait() [ 671.543168] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 671.543168] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] return self._exit_event.wait() [ 671.543168] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 671.543168] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] current.throw(*self._exc) [ 671.543168] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 671.543168] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] result = function(*args, **kwargs) [ 671.543554] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 671.543554] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] return func(*args, **kwargs) [ 671.543554] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 671.543554] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] raise e [ 671.543554] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 671.543554] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] nwinfo = self.network_api.allocate_for_instance( [ 671.543554] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 671.543554] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] created_port_ids = self._update_ports_for_instance( [ 671.543554] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 671.543554] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] with excutils.save_and_reraise_exception(): [ 671.543554] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.543554] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] self.force_reraise() [ 671.543554] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.543968] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] raise self.value [ 671.543968] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 671.543968] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] updated_port = self._update_port( [ 671.543968] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 671.543968] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] _ensure_no_port_binding_failure(port) [ 671.543968] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 671.543968] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] raise exception.PortBindingFailed(port_id=port['id']) [ 671.543968] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] nova.exception.PortBindingFailed: Binding failed for port 0b0dcc76-2a8c-4f06-b32e-22e39f072235, please check neutron logs for more information. [ 671.543968] env[63021]: ERROR nova.compute.manager [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] [ 671.543968] env[63021]: DEBUG nova.compute.utils [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Binding failed for port 0b0dcc76-2a8c-4f06-b32e-22e39f072235, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 671.544612] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 18.447s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 671.546417] env[63021]: DEBUG nova.compute.manager [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Build of instance 7da1d0f4-5dbe-4243-ac87-e476d5870882 was re-scheduled: Binding failed for port 0b0dcc76-2a8c-4f06-b32e-22e39f072235, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 671.546662] env[63021]: DEBUG nova.compute.manager [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 671.546935] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Acquiring lock "refresh_cache-7da1d0f4-5dbe-4243-ac87-e476d5870882" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.547028] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Acquired lock "refresh_cache-7da1d0f4-5dbe-4243-ac87-e476d5870882" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.547187] env[63021]: DEBUG nova.network.neutron [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 671.740610] env[63021]: INFO nova.compute.manager [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c] Took 1.02 seconds to deallocate network for instance. [ 671.991279] env[63021]: DEBUG oslo_vmware.api [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Task: {'id': task-1293586, 'name': PowerOnVM_Task, 'duration_secs': 0.409657} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.991549] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 671.991770] env[63021]: INFO nova.compute.manager [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Took 4.83 seconds to spawn the instance on the hypervisor. [ 671.991953] env[63021]: DEBUG nova.compute.manager [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 671.992265] env[63021]: DEBUG nova.network.neutron [-] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.993797] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1be0412-addb-49b5-8d30-b01e7c47373b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.066569] env[63021]: DEBUG nova.network.neutron [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.140912] env[63021]: DEBUG nova.network.neutron [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.496983] env[63021]: INFO nova.compute.manager [-] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Took 1.02 seconds to deallocate network for instance. [ 672.499914] env[63021]: DEBUG nova.compute.claims [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 672.500103] env[63021]: DEBUG oslo_concurrency.lockutils [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 672.510492] env[63021]: INFO nova.compute.manager [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Took 28.86 seconds to build instance. [ 672.643253] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Releasing lock "refresh_cache-7da1d0f4-5dbe-4243-ac87-e476d5870882" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.643489] env[63021]: DEBUG nova.compute.manager [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 672.643665] env[63021]: DEBUG nova.compute.manager [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 672.643919] env[63021]: DEBUG nova.network.neutron [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 672.659414] env[63021]: DEBUG nova.network.neutron [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.766908] env[63021]: INFO nova.scheduler.client.report [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Deleted allocations for instance 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c [ 673.012920] env[63021]: DEBUG oslo_concurrency.lockutils [None req-26aaee0d-bc6f-4583-ad42-8f54012aad18 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Lock "77fd1199-33b5-4040-a0ba-dd881da1224c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.512s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.075298] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 673.161708] env[63021]: DEBUG nova.network.neutron [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.278624] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5fa6f759-9da1-4c16-b1ca-9c81df61e946 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "12b2ab52-bbfb-47f4-b5ec-5c10bfb97b0c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 104.193s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.515563] env[63021]: DEBUG nova.compute.manager [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 673.578448] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 7da1d0f4-5dbe-4243-ac87-e476d5870882 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 673.578614] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance c76dfad0-2c67-4fe4-8c86-04ece5c80ea8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 673.578740] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 1aaa84d5-c5d5-4321-a98f-da261de5d364 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 673.578860] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance f5f546f8-291f-430c-95fb-c78507b0087f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 673.579108] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 673.579256] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 39f51f90-a5a5-4479-881c-b5b23a3bf01a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 673.579372] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 77fd1199-33b5-4040-a0ba-dd881da1224c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 673.579484] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 4e6dcd18-4814-4cf1-966b-7d853858e650 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 673.664299] env[63021]: INFO nova.compute.manager [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: 7da1d0f4-5dbe-4243-ac87-e476d5870882] Took 1.02 seconds to deallocate network for instance. [ 673.782042] env[63021]: DEBUG nova.compute.manager [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 673.852308] env[63021]: DEBUG nova.compute.manager [None req-3bf4c9d7-4e03-49bd-bda3-79d59de3aada tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 673.853287] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56f5f4d4-1606-4497-be19-303887de6e75 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.913897] env[63021]: DEBUG oslo_concurrency.lockutils [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Acquiring lock "77fd1199-33b5-4040-a0ba-dd881da1224c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.914313] env[63021]: DEBUG oslo_concurrency.lockutils [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Lock "77fd1199-33b5-4040-a0ba-dd881da1224c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.914526] env[63021]: DEBUG oslo_concurrency.lockutils [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Acquiring lock "77fd1199-33b5-4040-a0ba-dd881da1224c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.914764] env[63021]: DEBUG oslo_concurrency.lockutils [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Lock "77fd1199-33b5-4040-a0ba-dd881da1224c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.915018] env[63021]: DEBUG oslo_concurrency.lockutils [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Lock "77fd1199-33b5-4040-a0ba-dd881da1224c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.917386] env[63021]: INFO nova.compute.manager [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Terminating instance [ 673.919276] env[63021]: DEBUG oslo_concurrency.lockutils [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Acquiring lock "refresh_cache-77fd1199-33b5-4040-a0ba-dd881da1224c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 673.919489] env[63021]: DEBUG oslo_concurrency.lockutils [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Acquired lock "refresh_cache-77fd1199-33b5-4040-a0ba-dd881da1224c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.919757] env[63021]: DEBUG nova.network.neutron [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 674.039143] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.084220] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 52b854c4-ccc8-4e38-8100-31ab06959939 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 674.302013] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.364605] env[63021]: INFO nova.compute.manager [None req-3bf4c9d7-4e03-49bd-bda3-79d59de3aada tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] instance snapshotting [ 674.365439] env[63021]: DEBUG nova.objects.instance [None req-3bf4c9d7-4e03-49bd-bda3-79d59de3aada tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Lazy-loading 'flavor' on Instance uuid 77fd1199-33b5-4040-a0ba-dd881da1224c {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 674.438217] env[63021]: DEBUG nova.network.neutron [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 674.488163] env[63021]: DEBUG nova.network.neutron [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.589164] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance aef64522-2dbb-4bc5-84b2-a0bf8a108574 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 674.615212] env[63021]: DEBUG oslo_concurrency.lockutils [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "ad2037c0-ed67-4cd1-b614-1733acfd7fd0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.615448] env[63021]: DEBUG oslo_concurrency.lockutils [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "ad2037c0-ed67-4cd1-b614-1733acfd7fd0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.691037] env[63021]: INFO nova.scheduler.client.report [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Deleted allocations for instance 7da1d0f4-5dbe-4243-ac87-e476d5870882 [ 674.872172] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7920755d-ddba-43d9-88c2-7014ac285c68 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.889573] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6733a4d2-e2a4-4eba-be83-e6d84cfc6894 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.990996] env[63021]: DEBUG oslo_concurrency.lockutils [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Releasing lock "refresh_cache-77fd1199-33b5-4040-a0ba-dd881da1224c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 674.991699] env[63021]: DEBUG nova.compute.manager [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 674.991975] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 674.993734] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4f4467f-eaaa-49ab-9991-33602afc0ea9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.003990] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 675.004331] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2e57335e-9a63-48e7-82c2-827584f9577e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.010724] env[63021]: DEBUG oslo_vmware.api [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Waiting for the task: (returnval){ [ 675.010724] env[63021]: value = "task-1293587" [ 675.010724] env[63021]: _type = "Task" [ 675.010724] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.018658] env[63021]: DEBUG oslo_vmware.api [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Task: {'id': task-1293587, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.092661] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance c6ed1f3b-e80e-4545-8338-8e3aead5ba16 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 675.199139] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d6519308-69b7-4fe9-a071-4adea99df7ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Lock "7da1d0f4-5dbe-4243-ac87-e476d5870882" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.178s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.399604] env[63021]: DEBUG nova.compute.manager [None req-3bf4c9d7-4e03-49bd-bda3-79d59de3aada tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Instance disappeared during snapshot {{(pid=63021) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 675.520452] env[63021]: DEBUG oslo_vmware.api [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Task: {'id': task-1293587, 'name': PowerOffVM_Task, 'duration_secs': 0.180235} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.520722] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 675.520884] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 675.521143] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a73e812a-e3e1-420b-8996-32d0a4564427 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.531186] env[63021]: DEBUG nova.compute.manager [None req-3bf4c9d7-4e03-49bd-bda3-79d59de3aada tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Found 0 images (rotation: 2) {{(pid=63021) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 675.548297] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 675.548481] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 675.548662] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Deleting the datastore file [datastore1] 77fd1199-33b5-4040-a0ba-dd881da1224c {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 675.548919] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0d326a44-4b42-481d-940b-b5940ca2fcd0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.555976] env[63021]: DEBUG oslo_vmware.api [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Waiting for the task: (returnval){ [ 675.555976] env[63021]: value = "task-1293589" [ 675.555976] env[63021]: _type = "Task" [ 675.555976] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.565780] env[63021]: DEBUG oslo_vmware.api [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Task: {'id': task-1293589, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.596152] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance d36a9ed2-6849-43fc-ab88-77468bde337f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 675.701408] env[63021]: DEBUG nova.compute.manager [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 676.065372] env[63021]: DEBUG oslo_vmware.api [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Task: {'id': task-1293589, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.188766} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.065636] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 676.065800] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 676.065965] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 676.066152] env[63021]: INFO nova.compute.manager [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Took 1.07 seconds to destroy the instance on the hypervisor. [ 676.066384] env[63021]: DEBUG oslo.service.loopingcall [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 676.066567] env[63021]: DEBUG nova.compute.manager [-] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 676.066661] env[63021]: DEBUG nova.network.neutron [-] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 676.081597] env[63021]: DEBUG nova.network.neutron [-] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 676.100035] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 476e1e73-7050-4e53-bff8-d5d9f0982f1a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 676.225554] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 676.584016] env[63021]: DEBUG nova.network.neutron [-] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.602060] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 28d70e1d-da2a-4c9c-b297-86d026ce42c3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 677.086685] env[63021]: INFO nova.compute.manager [-] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Took 1.02 seconds to deallocate network for instance. [ 677.104630] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 03aec3ed-f338-41b5-a396-e3d693edbb8c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 677.593669] env[63021]: DEBUG oslo_concurrency.lockutils [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.607550] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 239d0087-462d-445a-a131-a23d4d15505f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 678.110886] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance c2523d70-12dd-444b-8839-b97378ef687e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 678.614167] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 26896531-07eb-4e44-8dd8-e5237bfd7c39 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 679.117742] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 274b001a-443f-463c-a657-652f64ba1de8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 679.622037] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 3660a926-7460-41f2-9ee4-2a5072701aea has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 680.124345] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance bc48ac80-36f6-4c97-9f95-29aed2efeae6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 680.628541] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 907540da-a701-477a-9b5c-9942b5d2d987 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 681.130619] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 06ef77f1-a196-499c-b639-8427b021a8c0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 681.633870] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 8ea2a693-0cd8-4f91-8bd8-b40265b9372e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 682.137132] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance d02adae0-724e-4dbe-a6c5-d476805d3d57 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 682.640204] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance f77da305-9cc9-415d-a808-c2136212c456 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 683.146950] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 9f3b0e67-24fc-4625-81f4-050055906c19 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 683.147254] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=63021) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 683.147387] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1856MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=63021) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 683.429842] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02ad49bf-77e2-432d-a633-9935aa11bfa9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.437315] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e73d14b1-0b18-4bf3-a8b4-08893605a052 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.466886] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec1aa876-ab8e-420c-a808-0b09efc3c67f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.474208] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b6f3172-7caf-46fc-aecc-207c58a75f2a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.486864] env[63021]: DEBUG nova.compute.provider_tree [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 683.989643] env[63021]: DEBUG nova.scheduler.client.report [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 684.496550] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63021) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 684.496550] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 12.952s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.496911] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.634s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.362159] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82a44633-b8dd-4ef8-ac15-7cce0b6930b7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.369564] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1224f3e6-9352-41f5-898f-6f2c646d3ca1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.401390] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7837fb84-7313-47a2-b601-030b1631dd3f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.409420] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f772ac24-6cc2-43dd-9241-576a40824b0d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.426486] env[63021]: DEBUG nova.compute.provider_tree [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 685.929733] env[63021]: DEBUG nova.scheduler.client.report [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 686.435360] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.938s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.436029] env[63021]: ERROR nova.compute.manager [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ed117e94-79ec-45ad-9605-16a99e4fb593, please check neutron logs for more information. [ 686.436029] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Traceback (most recent call last): [ 686.436029] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 686.436029] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] self.driver.spawn(context, instance, image_meta, [ 686.436029] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 686.436029] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 686.436029] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 686.436029] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] vm_ref = self.build_virtual_machine(instance, [ 686.436029] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 686.436029] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] vif_infos = vmwarevif.get_vif_info(self._session, [ 686.436029] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 686.436359] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] for vif in network_info: [ 686.436359] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 686.436359] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] return self._sync_wrapper(fn, *args, **kwargs) [ 686.436359] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 686.436359] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] self.wait() [ 686.436359] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 686.436359] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] self[:] = self._gt.wait() [ 686.436359] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 686.436359] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] return self._exit_event.wait() [ 686.436359] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 686.436359] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] current.throw(*self._exc) [ 686.436359] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 686.436359] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] result = function(*args, **kwargs) [ 686.436728] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 686.436728] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] return func(*args, **kwargs) [ 686.436728] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 686.436728] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] raise e [ 686.436728] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 686.436728] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] nwinfo = self.network_api.allocate_for_instance( [ 686.436728] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 686.436728] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] created_port_ids = self._update_ports_for_instance( [ 686.436728] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 686.436728] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] with excutils.save_and_reraise_exception(): [ 686.436728] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 686.436728] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] self.force_reraise() [ 686.436728] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 686.437101] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] raise self.value [ 686.437101] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 686.437101] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] updated_port = self._update_port( [ 686.437101] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 686.437101] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] _ensure_no_port_binding_failure(port) [ 686.437101] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 686.437101] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] raise exception.PortBindingFailed(port_id=port['id']) [ 686.437101] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] nova.exception.PortBindingFailed: Binding failed for port ed117e94-79ec-45ad-9605-16a99e4fb593, please check neutron logs for more information. [ 686.437101] env[63021]: ERROR nova.compute.manager [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] [ 686.437101] env[63021]: DEBUG nova.compute.utils [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Binding failed for port ed117e94-79ec-45ad-9605-16a99e4fb593, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 686.438139] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.802s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.441770] env[63021]: DEBUG nova.compute.manager [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Build of instance c76dfad0-2c67-4fe4-8c86-04ece5c80ea8 was re-scheduled: Binding failed for port ed117e94-79ec-45ad-9605-16a99e4fb593, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 686.442393] env[63021]: DEBUG nova.compute.manager [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 686.442621] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Acquiring lock "refresh_cache-c76dfad0-2c67-4fe4-8c86-04ece5c80ea8" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 686.442792] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Acquired lock "refresh_cache-c76dfad0-2c67-4fe4-8c86-04ece5c80ea8" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.442938] env[63021]: DEBUG nova.network.neutron [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 686.964116] env[63021]: DEBUG nova.network.neutron [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 687.040674] env[63021]: DEBUG nova.network.neutron [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.242217] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e90c45f6-4518-422c-8f0c-3335223cdfd3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.249824] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5409786d-dd60-41c4-bccd-4e528ded0f5a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.281776] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd73087a-c983-411a-b91b-2a996760ba8c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.289439] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e333bd9a-8e00-4aea-a7b7-2fb09f959ab1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.302387] env[63021]: DEBUG nova.compute.provider_tree [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 687.543618] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Releasing lock "refresh_cache-c76dfad0-2c67-4fe4-8c86-04ece5c80ea8" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.543868] env[63021]: DEBUG nova.compute.manager [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 687.544100] env[63021]: DEBUG nova.compute.manager [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 687.544245] env[63021]: DEBUG nova.network.neutron [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 687.558789] env[63021]: DEBUG nova.network.neutron [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 687.805554] env[63021]: DEBUG nova.scheduler.client.report [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 688.064297] env[63021]: DEBUG nova.network.neutron [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.311366] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.873s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.312194] env[63021]: ERROR nova.compute.manager [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2f32ab2a-287e-403b-b713-052e29aea30e, please check neutron logs for more information. [ 688.312194] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Traceback (most recent call last): [ 688.312194] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 688.312194] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] self.driver.spawn(context, instance, image_meta, [ 688.312194] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 688.312194] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] self._vmops.spawn(context, instance, image_meta, injected_files, [ 688.312194] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 688.312194] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] vm_ref = self.build_virtual_machine(instance, [ 688.312194] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 688.312194] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] vif_infos = vmwarevif.get_vif_info(self._session, [ 688.312194] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 688.312616] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] for vif in network_info: [ 688.312616] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 688.312616] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] return self._sync_wrapper(fn, *args, **kwargs) [ 688.312616] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 688.312616] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] self.wait() [ 688.312616] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 688.312616] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] self[:] = self._gt.wait() [ 688.312616] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 688.312616] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] return self._exit_event.wait() [ 688.312616] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 688.312616] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] result = hub.switch() [ 688.312616] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 688.312616] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] return self.greenlet.switch() [ 688.312985] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 688.312985] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] result = function(*args, **kwargs) [ 688.312985] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 688.312985] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] return func(*args, **kwargs) [ 688.312985] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 688.312985] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] raise e [ 688.312985] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 688.312985] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] nwinfo = self.network_api.allocate_for_instance( [ 688.312985] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 688.312985] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] created_port_ids = self._update_ports_for_instance( [ 688.312985] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 688.312985] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] with excutils.save_and_reraise_exception(): [ 688.312985] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 688.313362] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] self.force_reraise() [ 688.313362] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 688.313362] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] raise self.value [ 688.313362] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 688.313362] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] updated_port = self._update_port( [ 688.313362] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 688.313362] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] _ensure_no_port_binding_failure(port) [ 688.313362] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 688.313362] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] raise exception.PortBindingFailed(port_id=port['id']) [ 688.313362] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] nova.exception.PortBindingFailed: Binding failed for port 2f32ab2a-287e-403b-b713-052e29aea30e, please check neutron logs for more information. [ 688.313362] env[63021]: ERROR nova.compute.manager [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] [ 688.313791] env[63021]: DEBUG nova.compute.utils [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Binding failed for port 2f32ab2a-287e-403b-b713-052e29aea30e, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 688.314287] env[63021]: DEBUG oslo_concurrency.lockutils [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.477s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.317982] env[63021]: DEBUG nova.compute.manager [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Build of instance 1aaa84d5-c5d5-4321-a98f-da261de5d364 was re-scheduled: Binding failed for port 2f32ab2a-287e-403b-b713-052e29aea30e, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 688.318434] env[63021]: DEBUG nova.compute.manager [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 688.318662] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Acquiring lock "refresh_cache-1aaa84d5-c5d5-4321-a98f-da261de5d364" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.318806] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Acquired lock "refresh_cache-1aaa84d5-c5d5-4321-a98f-da261de5d364" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.318964] env[63021]: DEBUG nova.network.neutron [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 688.566791] env[63021]: INFO nova.compute.manager [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] [instance: c76dfad0-2c67-4fe4-8c86-04ece5c80ea8] Took 1.02 seconds to deallocate network for instance. [ 688.838901] env[63021]: DEBUG nova.network.neutron [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 688.909157] env[63021]: DEBUG nova.network.neutron [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.134351] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a772c095-bcf2-436d-b258-34b22d195ad2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.142899] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cae6f49-d991-4dfb-b0da-b3bde4a2bcb4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.177193] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-652273d3-8f81-4c49-b0b5-fdfb987d244e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.185089] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8471cc5-0578-4cdf-ba92-b9796af5e38f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.198125] env[63021]: DEBUG nova.compute.provider_tree [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 689.412043] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Releasing lock "refresh_cache-1aaa84d5-c5d5-4321-a98f-da261de5d364" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.412315] env[63021]: DEBUG nova.compute.manager [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 689.412505] env[63021]: DEBUG nova.compute.manager [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 689.412668] env[63021]: DEBUG nova.network.neutron [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 689.427900] env[63021]: DEBUG nova.network.neutron [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 689.602162] env[63021]: INFO nova.scheduler.client.report [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Deleted allocations for instance c76dfad0-2c67-4fe4-8c86-04ece5c80ea8 [ 689.701409] env[63021]: DEBUG nova.scheduler.client.report [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 689.930177] env[63021]: DEBUG nova.network.neutron [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.110577] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fcda8653-fd78-4013-ae77-6d23525985ec tempest-ServersAdminTestJSON-691789307 tempest-ServersAdminTestJSON-691789307-project-member] Lock "c76dfad0-2c67-4fe4-8c86-04ece5c80ea8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 118.966s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.206056] env[63021]: DEBUG oslo_concurrency.lockutils [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.892s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.206733] env[63021]: ERROR nova.compute.manager [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 45c257ff-e39a-4ca8-88b1-bc84407ea2d8, please check neutron logs for more information. [ 690.206733] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Traceback (most recent call last): [ 690.206733] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 690.206733] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] self.driver.spawn(context, instance, image_meta, [ 690.206733] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 690.206733] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 690.206733] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 690.206733] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] vm_ref = self.build_virtual_machine(instance, [ 690.206733] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 690.206733] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] vif_infos = vmwarevif.get_vif_info(self._session, [ 690.206733] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 690.207101] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] for vif in network_info: [ 690.207101] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 690.207101] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] return self._sync_wrapper(fn, *args, **kwargs) [ 690.207101] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 690.207101] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] self.wait() [ 690.207101] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 690.207101] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] self[:] = self._gt.wait() [ 690.207101] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 690.207101] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] return self._exit_event.wait() [ 690.207101] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 690.207101] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] current.throw(*self._exc) [ 690.207101] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 690.207101] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] result = function(*args, **kwargs) [ 690.207468] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 690.207468] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] return func(*args, **kwargs) [ 690.207468] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 690.207468] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] raise e [ 690.207468] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 690.207468] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] nwinfo = self.network_api.allocate_for_instance( [ 690.207468] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 690.207468] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] created_port_ids = self._update_ports_for_instance( [ 690.207468] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 690.207468] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] with excutils.save_and_reraise_exception(): [ 690.207468] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 690.207468] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] self.force_reraise() [ 690.207468] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 690.207828] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] raise self.value [ 690.207828] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 690.207828] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] updated_port = self._update_port( [ 690.207828] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 690.207828] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] _ensure_no_port_binding_failure(port) [ 690.207828] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 690.207828] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] raise exception.PortBindingFailed(port_id=port['id']) [ 690.207828] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] nova.exception.PortBindingFailed: Binding failed for port 45c257ff-e39a-4ca8-88b1-bc84407ea2d8, please check neutron logs for more information. [ 690.207828] env[63021]: ERROR nova.compute.manager [instance: f5f546f8-291f-430c-95fb-c78507b0087f] [ 690.207828] env[63021]: DEBUG nova.compute.utils [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Binding failed for port 45c257ff-e39a-4ca8-88b1-bc84407ea2d8, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 690.208718] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.938s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 690.210384] env[63021]: INFO nova.compute.claims [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 690.213231] env[63021]: DEBUG nova.compute.manager [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Build of instance f5f546f8-291f-430c-95fb-c78507b0087f was re-scheduled: Binding failed for port 45c257ff-e39a-4ca8-88b1-bc84407ea2d8, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 690.213633] env[63021]: DEBUG nova.compute.manager [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 690.213902] env[63021]: DEBUG oslo_concurrency.lockutils [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Acquiring lock "refresh_cache-f5f546f8-291f-430c-95fb-c78507b0087f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.214052] env[63021]: DEBUG oslo_concurrency.lockutils [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Acquired lock "refresh_cache-f5f546f8-291f-430c-95fb-c78507b0087f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.214228] env[63021]: DEBUG nova.network.neutron [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 690.433697] env[63021]: INFO nova.compute.manager [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: 1aaa84d5-c5d5-4321-a98f-da261de5d364] Took 1.02 seconds to deallocate network for instance. [ 690.613253] env[63021]: DEBUG nova.compute.manager [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 690.737452] env[63021]: DEBUG nova.network.neutron [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 690.780695] env[63021]: DEBUG nova.network.neutron [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.135385] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.286016] env[63021]: DEBUG oslo_concurrency.lockutils [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Releasing lock "refresh_cache-f5f546f8-291f-430c-95fb-c78507b0087f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.286016] env[63021]: DEBUG nova.compute.manager [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 691.286016] env[63021]: DEBUG nova.compute.manager [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 691.286016] env[63021]: DEBUG nova.network.neutron [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 691.312426] env[63021]: DEBUG nova.network.neutron [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 691.459571] env[63021]: INFO nova.scheduler.client.report [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Deleted allocations for instance 1aaa84d5-c5d5-4321-a98f-da261de5d364 [ 691.570814] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57181085-1f61-41f9-86e8-8f89cbb03050 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.578881] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e045856-0719-45fe-8629-efd8dd529c48 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.612934] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b35f969-25fb-4201-ac06-0896d08be816 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.621138] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e70d690-a8c5-4c5a-9e64-7332737e112f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.640895] env[63021]: DEBUG nova.compute.provider_tree [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 691.815701] env[63021]: DEBUG nova.network.neutron [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.974115] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c773de73-25d8-4ca7-a2e8-171feae652ac tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Lock "1aaa84d5-c5d5-4321-a98f-da261de5d364" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.952s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.145268] env[63021]: DEBUG nova.scheduler.client.report [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 692.318377] env[63021]: INFO nova.compute.manager [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] [instance: f5f546f8-291f-430c-95fb-c78507b0087f] Took 1.03 seconds to deallocate network for instance. [ 692.478289] env[63021]: DEBUG nova.compute.manager [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 692.654648] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.445s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.654648] env[63021]: DEBUG nova.compute.manager [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 692.658122] env[63021]: DEBUG oslo_concurrency.lockutils [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.223s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.005152] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.162592] env[63021]: DEBUG nova.compute.utils [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 693.164247] env[63021]: DEBUG nova.compute.manager [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 693.164509] env[63021]: DEBUG nova.network.neutron [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 693.208719] env[63021]: DEBUG nova.policy [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6ea19453dbdb4251a93dde2576aeff64', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '58693d1e991249f28914a39761cc2627', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 693.356227] env[63021]: INFO nova.scheduler.client.report [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Deleted allocations for instance f5f546f8-291f-430c-95fb-c78507b0087f [ 693.501649] env[63021]: DEBUG nova.network.neutron [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Successfully created port: 670c7666-bae2-4f9f-b75c-9a31a67a5b44 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 693.525175] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c493577-4191-40d9-966f-6d16e80759d8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.534542] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73ad973e-2f8b-4965-a898-43da6ede985a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.572241] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd814ece-dab0-430f-8683-9fa1163a566e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.580588] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a694d0aa-b2e2-4304-999f-4f1a0f92d11f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.594582] env[63021]: DEBUG nova.compute.provider_tree [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 693.670214] env[63021]: DEBUG nova.compute.manager [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 693.867810] env[63021]: DEBUG oslo_concurrency.lockutils [None req-32f638be-09ce-4c6d-a42f-c546f2aa9382 tempest-MigrationsAdminTest-1048602009 tempest-MigrationsAdminTest-1048602009-project-member] Lock "f5f546f8-291f-430c-95fb-c78507b0087f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.066s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.099022] env[63021]: DEBUG nova.scheduler.client.report [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 694.370697] env[63021]: DEBUG nova.compute.manager [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 694.449937] env[63021]: DEBUG nova.compute.manager [req-30f30855-5e2a-48e7-942a-223e7dadc26e req-a628eb18-16b8-4957-907d-99a3d94f5983 service nova] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Received event network-changed-670c7666-bae2-4f9f-b75c-9a31a67a5b44 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 694.450179] env[63021]: DEBUG nova.compute.manager [req-30f30855-5e2a-48e7-942a-223e7dadc26e req-a628eb18-16b8-4957-907d-99a3d94f5983 service nova] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Refreshing instance network info cache due to event network-changed-670c7666-bae2-4f9f-b75c-9a31a67a5b44. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 694.450401] env[63021]: DEBUG oslo_concurrency.lockutils [req-30f30855-5e2a-48e7-942a-223e7dadc26e req-a628eb18-16b8-4957-907d-99a3d94f5983 service nova] Acquiring lock "refresh_cache-52b854c4-ccc8-4e38-8100-31ab06959939" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.450541] env[63021]: DEBUG oslo_concurrency.lockutils [req-30f30855-5e2a-48e7-942a-223e7dadc26e req-a628eb18-16b8-4957-907d-99a3d94f5983 service nova] Acquired lock "refresh_cache-52b854c4-ccc8-4e38-8100-31ab06959939" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.450692] env[63021]: DEBUG nova.network.neutron [req-30f30855-5e2a-48e7-942a-223e7dadc26e req-a628eb18-16b8-4957-907d-99a3d94f5983 service nova] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Refreshing network info cache for port 670c7666-bae2-4f9f-b75c-9a31a67a5b44 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 694.508334] env[63021]: ERROR nova.compute.manager [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 670c7666-bae2-4f9f-b75c-9a31a67a5b44, please check neutron logs for more information. [ 694.508334] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 694.508334] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.508334] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 694.508334] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 694.508334] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 694.508334] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 694.508334] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 694.508334] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.508334] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 694.508334] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.508334] env[63021]: ERROR nova.compute.manager raise self.value [ 694.508334] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 694.508334] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 694.508334] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.508334] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 694.509209] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.509209] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 694.509209] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 670c7666-bae2-4f9f-b75c-9a31a67a5b44, please check neutron logs for more information. [ 694.509209] env[63021]: ERROR nova.compute.manager [ 694.509209] env[63021]: Traceback (most recent call last): [ 694.509209] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 694.509209] env[63021]: listener.cb(fileno) [ 694.509209] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.509209] env[63021]: result = function(*args, **kwargs) [ 694.509209] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 694.509209] env[63021]: return func(*args, **kwargs) [ 694.509209] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 694.509209] env[63021]: raise e [ 694.509209] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.509209] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 694.509209] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 694.509209] env[63021]: created_port_ids = self._update_ports_for_instance( [ 694.509209] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 694.509209] env[63021]: with excutils.save_and_reraise_exception(): [ 694.509209] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.509209] env[63021]: self.force_reraise() [ 694.509209] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.509209] env[63021]: raise self.value [ 694.509209] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 694.509209] env[63021]: updated_port = self._update_port( [ 694.509209] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.509209] env[63021]: _ensure_no_port_binding_failure(port) [ 694.509209] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.509209] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 694.510502] env[63021]: nova.exception.PortBindingFailed: Binding failed for port 670c7666-bae2-4f9f-b75c-9a31a67a5b44, please check neutron logs for more information. [ 694.510502] env[63021]: Removing descriptor: 18 [ 694.605267] env[63021]: DEBUG oslo_concurrency.lockutils [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.946s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.605267] env[63021]: ERROR nova.compute.manager [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 31850b77-7688-482d-91fb-d74fdd83e293, please check neutron logs for more information. [ 694.605267] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Traceback (most recent call last): [ 694.605267] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 694.605267] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] self.driver.spawn(context, instance, image_meta, [ 694.605267] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 694.605267] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 694.605267] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 694.605267] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] vm_ref = self.build_virtual_machine(instance, [ 694.605664] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 694.605664] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] vif_infos = vmwarevif.get_vif_info(self._session, [ 694.605664] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 694.605664] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] for vif in network_info: [ 694.605664] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 694.605664] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] return self._sync_wrapper(fn, *args, **kwargs) [ 694.605664] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 694.605664] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] self.wait() [ 694.605664] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 694.605664] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] self[:] = self._gt.wait() [ 694.605664] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 694.605664] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] return self._exit_event.wait() [ 694.605664] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 694.606223] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] result = hub.switch() [ 694.606223] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 694.606223] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] return self.greenlet.switch() [ 694.606223] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.606223] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] result = function(*args, **kwargs) [ 694.606223] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 694.606223] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] return func(*args, **kwargs) [ 694.606223] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 694.606223] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] raise e [ 694.606223] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.606223] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] nwinfo = self.network_api.allocate_for_instance( [ 694.606223] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 694.606223] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] created_port_ids = self._update_ports_for_instance( [ 694.606543] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 694.606543] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] with excutils.save_and_reraise_exception(): [ 694.606543] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.606543] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] self.force_reraise() [ 694.606543] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.606543] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] raise self.value [ 694.606543] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 694.606543] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] updated_port = self._update_port( [ 694.606543] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.606543] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] _ensure_no_port_binding_failure(port) [ 694.606543] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.606543] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] raise exception.PortBindingFailed(port_id=port['id']) [ 694.606820] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] nova.exception.PortBindingFailed: Binding failed for port 31850b77-7688-482d-91fb-d74fdd83e293, please check neutron logs for more information. [ 694.606820] env[63021]: ERROR nova.compute.manager [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] [ 694.606820] env[63021]: DEBUG nova.compute.utils [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Binding failed for port 31850b77-7688-482d-91fb-d74fdd83e293, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 694.609623] env[63021]: DEBUG nova.compute.manager [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Build of instance 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6 was re-scheduled: Binding failed for port 31850b77-7688-482d-91fb-d74fdd83e293, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 694.609623] env[63021]: DEBUG nova.compute.manager [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 694.609623] env[63021]: DEBUG oslo_concurrency.lockutils [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Acquiring lock "refresh_cache-05caabab-8c7d-4d27-8eb3-e7e1c2947aa6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.609623] env[63021]: DEBUG oslo_concurrency.lockutils [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Acquired lock "refresh_cache-05caabab-8c7d-4d27-8eb3-e7e1c2947aa6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.609826] env[63021]: DEBUG nova.network.neutron [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 694.611033] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.599s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.683661] env[63021]: DEBUG nova.compute.manager [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 694.714054] env[63021]: DEBUG nova.virt.hardware [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 694.714054] env[63021]: DEBUG nova.virt.hardware [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 694.714054] env[63021]: DEBUG nova.virt.hardware [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 694.714426] env[63021]: DEBUG nova.virt.hardware [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 694.714426] env[63021]: DEBUG nova.virt.hardware [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 694.714426] env[63021]: DEBUG nova.virt.hardware [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 694.714426] env[63021]: DEBUG nova.virt.hardware [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 694.714426] env[63021]: DEBUG nova.virt.hardware [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 694.714563] env[63021]: DEBUG nova.virt.hardware [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 694.714563] env[63021]: DEBUG nova.virt.hardware [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 694.714563] env[63021]: DEBUG nova.virt.hardware [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 694.716025] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3af9e62d-f077-4324-ba8e-493eb3c4b8c3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.725850] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-707aa04d-09d2-4d26-a7cb-8741538d2f33 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.739845] env[63021]: ERROR nova.compute.manager [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 670c7666-bae2-4f9f-b75c-9a31a67a5b44, please check neutron logs for more information. [ 694.739845] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Traceback (most recent call last): [ 694.739845] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 694.739845] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] yield resources [ 694.739845] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 694.739845] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] self.driver.spawn(context, instance, image_meta, [ 694.739845] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 694.739845] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] self._vmops.spawn(context, instance, image_meta, injected_files, [ 694.739845] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 694.739845] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] vm_ref = self.build_virtual_machine(instance, [ 694.739845] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 694.740219] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] vif_infos = vmwarevif.get_vif_info(self._session, [ 694.740219] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 694.740219] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] for vif in network_info: [ 694.740219] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 694.740219] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] return self._sync_wrapper(fn, *args, **kwargs) [ 694.740219] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 694.740219] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] self.wait() [ 694.740219] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 694.740219] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] self[:] = self._gt.wait() [ 694.740219] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 694.740219] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] return self._exit_event.wait() [ 694.740219] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 694.740219] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] current.throw(*self._exc) [ 694.740581] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.740581] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] result = function(*args, **kwargs) [ 694.740581] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 694.740581] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] return func(*args, **kwargs) [ 694.740581] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 694.740581] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] raise e [ 694.740581] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.740581] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] nwinfo = self.network_api.allocate_for_instance( [ 694.740581] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 694.740581] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] created_port_ids = self._update_ports_for_instance( [ 694.740581] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 694.740581] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] with excutils.save_and_reraise_exception(): [ 694.740581] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.740896] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] self.force_reraise() [ 694.740896] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.740896] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] raise self.value [ 694.740896] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 694.740896] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] updated_port = self._update_port( [ 694.740896] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.740896] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] _ensure_no_port_binding_failure(port) [ 694.740896] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.740896] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] raise exception.PortBindingFailed(port_id=port['id']) [ 694.740896] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] nova.exception.PortBindingFailed: Binding failed for port 670c7666-bae2-4f9f-b75c-9a31a67a5b44, please check neutron logs for more information. [ 694.740896] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] [ 694.740896] env[63021]: INFO nova.compute.manager [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Terminating instance [ 694.742148] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "refresh_cache-52b854c4-ccc8-4e38-8100-31ab06959939" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.897785] env[63021]: DEBUG oslo_concurrency.lockutils [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.991175] env[63021]: DEBUG nova.network.neutron [req-30f30855-5e2a-48e7-942a-223e7dadc26e req-a628eb18-16b8-4957-907d-99a3d94f5983 service nova] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 695.097914] env[63021]: DEBUG nova.network.neutron [req-30f30855-5e2a-48e7-942a-223e7dadc26e req-a628eb18-16b8-4957-907d-99a3d94f5983 service nova] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.129772] env[63021]: DEBUG nova.network.neutron [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 695.192222] env[63021]: DEBUG nova.network.neutron [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.463961] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8f6b193-3d40-4adb-840a-658f72a71393 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.472066] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc98b33b-585e-4dee-a404-687cac23db64 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.501868] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c062111-cdc8-4135-b6df-fde6cf5689f6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.509324] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-322eab7c-ecaa-4650-a133-ecd271f1f803 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.522440] env[63021]: DEBUG nova.compute.provider_tree [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 695.603654] env[63021]: DEBUG oslo_concurrency.lockutils [req-30f30855-5e2a-48e7-942a-223e7dadc26e req-a628eb18-16b8-4957-907d-99a3d94f5983 service nova] Releasing lock "refresh_cache-52b854c4-ccc8-4e38-8100-31ab06959939" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.604381] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquired lock "refresh_cache-52b854c4-ccc8-4e38-8100-31ab06959939" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.604381] env[63021]: DEBUG nova.network.neutron [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 695.697398] env[63021]: DEBUG oslo_concurrency.lockutils [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Releasing lock "refresh_cache-05caabab-8c7d-4d27-8eb3-e7e1c2947aa6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.697398] env[63021]: DEBUG nova.compute.manager [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 695.697398] env[63021]: DEBUG nova.compute.manager [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 695.697398] env[63021]: DEBUG nova.network.neutron [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 695.710898] env[63021]: DEBUG nova.network.neutron [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.025583] env[63021]: DEBUG nova.scheduler.client.report [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 696.128209] env[63021]: DEBUG nova.network.neutron [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.218899] env[63021]: DEBUG nova.network.neutron [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.233295] env[63021]: DEBUG nova.network.neutron [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.477841] env[63021]: DEBUG nova.compute.manager [req-2bcc87d7-1101-4d9e-8010-47bbaadd3cc2 req-2884fe85-e748-4606-b030-a3195a668edf service nova] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Received event network-vif-deleted-670c7666-bae2-4f9f-b75c-9a31a67a5b44 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 696.530801] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.920s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 696.531525] env[63021]: ERROR nova.compute.manager [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port da5bf4a2-c61a-4810-8e49-677fa4a48055, please check neutron logs for more information. [ 696.531525] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Traceback (most recent call last): [ 696.531525] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 696.531525] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] self.driver.spawn(context, instance, image_meta, [ 696.531525] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 696.531525] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 696.531525] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 696.531525] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] vm_ref = self.build_virtual_machine(instance, [ 696.531525] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 696.531525] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] vif_infos = vmwarevif.get_vif_info(self._session, [ 696.531525] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 696.532031] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] for vif in network_info: [ 696.532031] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 696.532031] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] return self._sync_wrapper(fn, *args, **kwargs) [ 696.532031] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 696.532031] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] self.wait() [ 696.532031] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 696.532031] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] self[:] = self._gt.wait() [ 696.532031] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 696.532031] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] return self._exit_event.wait() [ 696.532031] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 696.532031] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] result = hub.switch() [ 696.532031] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 696.532031] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] return self.greenlet.switch() [ 696.532446] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 696.532446] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] result = function(*args, **kwargs) [ 696.532446] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 696.532446] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] return func(*args, **kwargs) [ 696.532446] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 696.532446] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] raise e [ 696.532446] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 696.532446] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] nwinfo = self.network_api.allocate_for_instance( [ 696.532446] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 696.532446] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] created_port_ids = self._update_ports_for_instance( [ 696.532446] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 696.532446] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] with excutils.save_and_reraise_exception(): [ 696.532446] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 696.532793] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] self.force_reraise() [ 696.532793] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 696.532793] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] raise self.value [ 696.532793] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 696.532793] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] updated_port = self._update_port( [ 696.532793] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 696.532793] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] _ensure_no_port_binding_failure(port) [ 696.532793] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 696.532793] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] raise exception.PortBindingFailed(port_id=port['id']) [ 696.532793] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] nova.exception.PortBindingFailed: Binding failed for port da5bf4a2-c61a-4810-8e49-677fa4a48055, please check neutron logs for more information. [ 696.532793] env[63021]: ERROR nova.compute.manager [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] [ 696.533461] env[63021]: DEBUG nova.compute.utils [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Binding failed for port da5bf4a2-c61a-4810-8e49-677fa4a48055, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 696.534905] env[63021]: DEBUG nova.compute.manager [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Build of instance 39f51f90-a5a5-4479-881c-b5b23a3bf01a was re-scheduled: Binding failed for port da5bf4a2-c61a-4810-8e49-677fa4a48055, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 696.535348] env[63021]: DEBUG nova.compute.manager [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 696.535660] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Acquiring lock "refresh_cache-39f51f90-a5a5-4479-881c-b5b23a3bf01a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.535737] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Acquired lock "refresh_cache-39f51f90-a5a5-4479-881c-b5b23a3bf01a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.535863] env[63021]: DEBUG nova.network.neutron [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 696.540190] env[63021]: DEBUG oslo_concurrency.lockutils [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.039s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 696.724393] env[63021]: INFO nova.compute.manager [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] [instance: 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6] Took 1.03 seconds to deallocate network for instance. [ 696.730545] env[63021]: DEBUG oslo_concurrency.lockutils [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Acquiring lock "f225812d-f6bf-4350-9151-9ab6c54cce5d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.731044] env[63021]: DEBUG oslo_concurrency.lockutils [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Lock "f225812d-f6bf-4350-9151-9ab6c54cce5d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 696.736223] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Releasing lock "refresh_cache-52b854c4-ccc8-4e38-8100-31ab06959939" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.736590] env[63021]: DEBUG nova.compute.manager [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 696.736765] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 696.737045] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-35fa7e9b-96ae-45a7-b9cc-0200a63c53ae {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.746558] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40e146a7-37d0-4421-a7e4-18d985be0ba3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.769088] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 52b854c4-ccc8-4e38-8100-31ab06959939 could not be found. [ 696.769292] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 696.769475] env[63021]: INFO nova.compute.manager [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Took 0.03 seconds to destroy the instance on the hypervisor. [ 696.769704] env[63021]: DEBUG oslo.service.loopingcall [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 696.769981] env[63021]: DEBUG nova.compute.manager [-] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 696.770123] env[63021]: DEBUG nova.network.neutron [-] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 696.794369] env[63021]: DEBUG nova.network.neutron [-] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 697.058824] env[63021]: DEBUG nova.network.neutron [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 697.191178] env[63021]: DEBUG nova.network.neutron [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.296470] env[63021]: DEBUG nova.network.neutron [-] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.353812] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23b20aff-9ff9-4efc-bc63-71cb44287e38 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.361499] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba388d40-f6d1-4ed8-bda0-f4fe582373a3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.392741] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82a148c7-95d5-4020-a8ff-170b8ea567c0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.400333] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1344c5b6-b36b-4f87-8581-0ecb636c9136 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.417928] env[63021]: DEBUG nova.compute.provider_tree [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 697.532525] env[63021]: DEBUG oslo_concurrency.lockutils [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Acquiring lock "9a6720ef-5307-44b7-b566-8d553c9b5384" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 697.533042] env[63021]: DEBUG oslo_concurrency.lockutils [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Lock "9a6720ef-5307-44b7-b566-8d553c9b5384" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 697.693116] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Releasing lock "refresh_cache-39f51f90-a5a5-4479-881c-b5b23a3bf01a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.696051] env[63021]: DEBUG nova.compute.manager [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 697.696051] env[63021]: DEBUG nova.compute.manager [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 697.696051] env[63021]: DEBUG nova.network.neutron [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 697.713608] env[63021]: DEBUG nova.network.neutron [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 697.757516] env[63021]: INFO nova.scheduler.client.report [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Deleted allocations for instance 05caabab-8c7d-4d27-8eb3-e7e1c2947aa6 [ 697.803885] env[63021]: INFO nova.compute.manager [-] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Took 1.03 seconds to deallocate network for instance. [ 697.807938] env[63021]: DEBUG nova.compute.claims [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 697.808132] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 697.921186] env[63021]: DEBUG nova.scheduler.client.report [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 698.215466] env[63021]: DEBUG nova.network.neutron [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.268051] env[63021]: DEBUG oslo_concurrency.lockutils [None req-31c145ec-a11b-4cb2-a0f3-48a9ca77eccc tempest-ServerExternalEventsTest-1415870946 tempest-ServerExternalEventsTest-1415870946-project-member] Lock "05caabab-8c7d-4d27-8eb3-e7e1c2947aa6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.773s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.428592] env[63021]: DEBUG oslo_concurrency.lockutils [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.889s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.429150] env[63021]: ERROR nova.compute.manager [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 559cf9af-5956-4bb5-8fec-4f748c72d59b, please check neutron logs for more information. [ 698.429150] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Traceback (most recent call last): [ 698.429150] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 698.429150] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] self.driver.spawn(context, instance, image_meta, [ 698.429150] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 698.429150] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] self._vmops.spawn(context, instance, image_meta, injected_files, [ 698.429150] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 698.429150] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] vm_ref = self.build_virtual_machine(instance, [ 698.429150] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 698.429150] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] vif_infos = vmwarevif.get_vif_info(self._session, [ 698.429150] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 698.429547] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] for vif in network_info: [ 698.429547] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 698.429547] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] return self._sync_wrapper(fn, *args, **kwargs) [ 698.429547] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 698.429547] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] self.wait() [ 698.429547] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 698.429547] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] self[:] = self._gt.wait() [ 698.429547] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 698.429547] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] return self._exit_event.wait() [ 698.429547] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 698.429547] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] current.throw(*self._exc) [ 698.429547] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 698.429547] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] result = function(*args, **kwargs) [ 698.431885] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 698.431885] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] return func(*args, **kwargs) [ 698.431885] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 698.431885] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] raise e [ 698.431885] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 698.431885] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] nwinfo = self.network_api.allocate_for_instance( [ 698.431885] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 698.431885] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] created_port_ids = self._update_ports_for_instance( [ 698.431885] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 698.431885] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] with excutils.save_and_reraise_exception(): [ 698.431885] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 698.431885] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] self.force_reraise() [ 698.431885] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 698.432536] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] raise self.value [ 698.432536] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 698.432536] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] updated_port = self._update_port( [ 698.432536] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 698.432536] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] _ensure_no_port_binding_failure(port) [ 698.432536] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 698.432536] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] raise exception.PortBindingFailed(port_id=port['id']) [ 698.432536] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] nova.exception.PortBindingFailed: Binding failed for port 559cf9af-5956-4bb5-8fec-4f748c72d59b, please check neutron logs for more information. [ 698.432536] env[63021]: ERROR nova.compute.manager [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] [ 698.432536] env[63021]: DEBUG nova.compute.utils [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Binding failed for port 559cf9af-5956-4bb5-8fec-4f748c72d59b, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 698.432794] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.392s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.432895] env[63021]: INFO nova.compute.claims [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 698.436572] env[63021]: DEBUG nova.compute.manager [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Build of instance 4e6dcd18-4814-4cf1-966b-7d853858e650 was re-scheduled: Binding failed for port 559cf9af-5956-4bb5-8fec-4f748c72d59b, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 698.436572] env[63021]: DEBUG nova.compute.manager [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 698.436572] env[63021]: DEBUG oslo_concurrency.lockutils [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Acquiring lock "refresh_cache-4e6dcd18-4814-4cf1-966b-7d853858e650" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.436572] env[63021]: DEBUG oslo_concurrency.lockutils [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Acquired lock "refresh_cache-4e6dcd18-4814-4cf1-966b-7d853858e650" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.436749] env[63021]: DEBUG nova.network.neutron [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 698.721182] env[63021]: INFO nova.compute.manager [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] [instance: 39f51f90-a5a5-4479-881c-b5b23a3bf01a] Took 1.03 seconds to deallocate network for instance. [ 698.770767] env[63021]: DEBUG nova.compute.manager [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 698.959096] env[63021]: DEBUG nova.network.neutron [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 699.061969] env[63021]: DEBUG nova.network.neutron [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.295762] env[63021]: DEBUG oslo_concurrency.lockutils [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 699.357419] env[63021]: DEBUG oslo_concurrency.lockutils [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Acquiring lock "929cf4cd-6f3c-4988-8f80-d889d3070e1e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 699.360103] env[63021]: DEBUG oslo_concurrency.lockutils [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Lock "929cf4cd-6f3c-4988-8f80-d889d3070e1e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.564633] env[63021]: DEBUG oslo_concurrency.lockutils [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Releasing lock "refresh_cache-4e6dcd18-4814-4cf1-966b-7d853858e650" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 699.565686] env[63021]: DEBUG nova.compute.manager [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 699.565686] env[63021]: DEBUG nova.compute.manager [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 699.565686] env[63021]: DEBUG nova.network.neutron [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 699.580707] env[63021]: DEBUG nova.network.neutron [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 699.751134] env[63021]: INFO nova.scheduler.client.report [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Deleted allocations for instance 39f51f90-a5a5-4479-881c-b5b23a3bf01a [ 699.781331] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3556fae9-f79f-4911-b5ac-1db8accad0c3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.795979] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2264b17-157a-475d-8f5b-3e44b2ac605c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.838251] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37daf5fa-f614-4853-b4fc-885648173a04 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.846260] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86b8ccb5-5e92-429a-ada0-52d511638571 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.859510] env[63021]: DEBUG nova.compute.provider_tree [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 700.085487] env[63021]: DEBUG nova.network.neutron [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.260410] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d636ca3b-73be-41ef-aa1f-4c958841dd52 tempest-AttachInterfacesUnderV243Test-398606289 tempest-AttachInterfacesUnderV243Test-398606289-project-member] Lock "39f51f90-a5a5-4479-881c-b5b23a3bf01a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 113.437s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 700.362364] env[63021]: DEBUG nova.scheduler.client.report [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 700.588449] env[63021]: INFO nova.compute.manager [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] [instance: 4e6dcd18-4814-4cf1-966b-7d853858e650] Took 1.02 seconds to deallocate network for instance. [ 700.766434] env[63021]: DEBUG nova.compute.manager [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 700.867370] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.436s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 700.867759] env[63021]: DEBUG nova.compute.manager [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 700.874023] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.569s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 700.874023] env[63021]: INFO nova.compute.claims [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 701.289996] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.373379] env[63021]: DEBUG nova.compute.utils [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 701.374897] env[63021]: DEBUG nova.compute.manager [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 701.376326] env[63021]: DEBUG nova.network.neutron [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 701.425931] env[63021]: DEBUG nova.policy [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '68e27cdf86b148bb99085f482afe51cb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9e3d1b4ec4ee4025912789c18e14b14b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 701.631041] env[63021]: INFO nova.scheduler.client.report [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Deleted allocations for instance 4e6dcd18-4814-4cf1-966b-7d853858e650 [ 701.678910] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Acquiring lock "7c6d7839-616d-41f5-a909-f8c626477a5e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.679178] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Lock "7c6d7839-616d-41f5-a909-f8c626477a5e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.720131] env[63021]: DEBUG nova.network.neutron [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Successfully created port: a1802ba1-ac36-4a83-a474-7c6e6253ca38 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 701.879979] env[63021]: DEBUG nova.compute.manager [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 702.142995] env[63021]: DEBUG oslo_concurrency.lockutils [None req-da585f1d-1321-4379-af8a-58482718fe72 tempest-ImagesNegativeTestJSON-372851194 tempest-ImagesNegativeTestJSON-372851194-project-member] Lock "4e6dcd18-4814-4cf1-966b-7d853858e650" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.869s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 702.220681] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a3cc168-7584-47b7-8561-dacb9f775160 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.229372] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49d93278-5c55-4c00-b247-6dbcbe12f5f9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.259667] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fda70c75-0d70-475e-a948-82ff53c97da0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.267260] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6549a16b-20c6-422d-9178-885c4d1c775b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.282674] env[63021]: DEBUG nova.compute.provider_tree [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 702.385702] env[63021]: INFO nova.virt.block_device [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Booting with volume ced55d5f-05a4-4276-8b51-00b63b5231c9 at /dev/sda [ 702.441459] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f4b0a91e-81cf-47f1-8fba-2fbfaa21311b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.450511] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a94eee28-2dee-4e5e-998c-ddff1432d9a2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.472355] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-51e428b3-439e-415c-913d-1e1230343d24 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.480074] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-497533f2-d4b9-4448-a6bc-a447ab1dc565 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.504205] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbce8a6d-545e-497c-b83e-b40bd061d240 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.510322] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d1e197b-b08a-4bab-858b-5e3a11ec2574 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.523361] env[63021]: DEBUG nova.virt.block_device [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Updating existing volume attachment record: f1130003-a77d-47f0-b845-9917467ff7b0 {{(pid=63021) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 702.541641] env[63021]: DEBUG nova.compute.manager [req-b399cd33-0e14-4964-a326-2fe0d84e916d req-7bf994d7-400e-44c5-af3d-63e1754d8272 service nova] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Received event network-changed-a1802ba1-ac36-4a83-a474-7c6e6253ca38 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 702.541745] env[63021]: DEBUG nova.compute.manager [req-b399cd33-0e14-4964-a326-2fe0d84e916d req-7bf994d7-400e-44c5-af3d-63e1754d8272 service nova] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Refreshing instance network info cache due to event network-changed-a1802ba1-ac36-4a83-a474-7c6e6253ca38. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 702.541956] env[63021]: DEBUG oslo_concurrency.lockutils [req-b399cd33-0e14-4964-a326-2fe0d84e916d req-7bf994d7-400e-44c5-af3d-63e1754d8272 service nova] Acquiring lock "refresh_cache-aef64522-2dbb-4bc5-84b2-a0bf8a108574" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.542138] env[63021]: DEBUG oslo_concurrency.lockutils [req-b399cd33-0e14-4964-a326-2fe0d84e916d req-7bf994d7-400e-44c5-af3d-63e1754d8272 service nova] Acquired lock "refresh_cache-aef64522-2dbb-4bc5-84b2-a0bf8a108574" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.542298] env[63021]: DEBUG nova.network.neutron [req-b399cd33-0e14-4964-a326-2fe0d84e916d req-7bf994d7-400e-44c5-af3d-63e1754d8272 service nova] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Refreshing network info cache for port a1802ba1-ac36-4a83-a474-7c6e6253ca38 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 702.646463] env[63021]: DEBUG nova.compute.manager [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 702.753568] env[63021]: ERROR nova.compute.manager [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a1802ba1-ac36-4a83-a474-7c6e6253ca38, please check neutron logs for more information. [ 702.753568] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 702.753568] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 702.753568] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 702.753568] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 702.753568] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 702.753568] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 702.753568] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 702.753568] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 702.753568] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 702.753568] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 702.753568] env[63021]: ERROR nova.compute.manager raise self.value [ 702.753568] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 702.753568] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 702.753568] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 702.753568] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 702.754202] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 702.754202] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 702.754202] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a1802ba1-ac36-4a83-a474-7c6e6253ca38, please check neutron logs for more information. [ 702.754202] env[63021]: ERROR nova.compute.manager [ 702.754202] env[63021]: Traceback (most recent call last): [ 702.754202] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 702.754202] env[63021]: listener.cb(fileno) [ 702.754202] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 702.754202] env[63021]: result = function(*args, **kwargs) [ 702.754202] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 702.754202] env[63021]: return func(*args, **kwargs) [ 702.754202] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 702.754202] env[63021]: raise e [ 702.754202] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 702.754202] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 702.754202] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 702.754202] env[63021]: created_port_ids = self._update_ports_for_instance( [ 702.754202] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 702.754202] env[63021]: with excutils.save_and_reraise_exception(): [ 702.754202] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 702.754202] env[63021]: self.force_reraise() [ 702.754202] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 702.754202] env[63021]: raise self.value [ 702.754202] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 702.754202] env[63021]: updated_port = self._update_port( [ 702.754202] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 702.754202] env[63021]: _ensure_no_port_binding_failure(port) [ 702.754202] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 702.754202] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 702.755032] env[63021]: nova.exception.PortBindingFailed: Binding failed for port a1802ba1-ac36-4a83-a474-7c6e6253ca38, please check neutron logs for more information. [ 702.755032] env[63021]: Removing descriptor: 18 [ 702.786035] env[63021]: DEBUG nova.scheduler.client.report [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 703.062087] env[63021]: DEBUG nova.network.neutron [req-b399cd33-0e14-4964-a326-2fe0d84e916d req-7bf994d7-400e-44c5-af3d-63e1754d8272 service nova] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.149229] env[63021]: DEBUG nova.network.neutron [req-b399cd33-0e14-4964-a326-2fe0d84e916d req-7bf994d7-400e-44c5-af3d-63e1754d8272 service nova] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.172054] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 703.291181] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.420s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.291815] env[63021]: DEBUG nova.compute.manager [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 703.296890] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.071s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.298585] env[63021]: INFO nova.compute.claims [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 703.653637] env[63021]: DEBUG oslo_concurrency.lockutils [req-b399cd33-0e14-4964-a326-2fe0d84e916d req-7bf994d7-400e-44c5-af3d-63e1754d8272 service nova] Releasing lock "refresh_cache-aef64522-2dbb-4bc5-84b2-a0bf8a108574" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.803175] env[63021]: DEBUG nova.compute.utils [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 703.808022] env[63021]: DEBUG nova.compute.manager [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 703.808022] env[63021]: DEBUG nova.network.neutron [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 703.889723] env[63021]: DEBUG nova.policy [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e2ce7bd37c324e76bad16423ba0e514a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4a0b1326882b40acb72aad7f6c6e2b0a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 704.310226] env[63021]: DEBUG nova.compute.manager [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 704.505451] env[63021]: DEBUG nova.network.neutron [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Successfully created port: ea5445cf-f96e-48c9-8ee4-e1e62207c32e {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 704.664035] env[63021]: DEBUG nova.compute.manager [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 704.664593] env[63021]: DEBUG nova.virt.hardware [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 704.664788] env[63021]: DEBUG nova.virt.hardware [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 704.664936] env[63021]: DEBUG nova.virt.hardware [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 704.665132] env[63021]: DEBUG nova.virt.hardware [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 704.665279] env[63021]: DEBUG nova.virt.hardware [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 704.665422] env[63021]: DEBUG nova.virt.hardware [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 704.665618] env[63021]: DEBUG nova.virt.hardware [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 704.665769] env[63021]: DEBUG nova.virt.hardware [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 704.665925] env[63021]: DEBUG nova.virt.hardware [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 704.667517] env[63021]: DEBUG nova.virt.hardware [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 704.667770] env[63021]: DEBUG nova.virt.hardware [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 704.674180] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5904f257-c7f9-4418-95b4-5707d8d93a78 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.686165] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c1b4362-024f-4d96-b2f0-1a67f53309ed {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.700886] env[63021]: ERROR nova.compute.manager [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a1802ba1-ac36-4a83-a474-7c6e6253ca38, please check neutron logs for more information. [ 704.700886] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Traceback (most recent call last): [ 704.700886] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 704.700886] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] yield resources [ 704.700886] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 704.700886] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] self.driver.spawn(context, instance, image_meta, [ 704.700886] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 704.700886] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] self._vmops.spawn(context, instance, image_meta, injected_files, [ 704.700886] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 704.700886] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] vm_ref = self.build_virtual_machine(instance, [ 704.700886] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 704.701333] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] vif_infos = vmwarevif.get_vif_info(self._session, [ 704.701333] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 704.701333] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] for vif in network_info: [ 704.701333] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 704.701333] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] return self._sync_wrapper(fn, *args, **kwargs) [ 704.701333] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 704.701333] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] self.wait() [ 704.701333] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 704.701333] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] self[:] = self._gt.wait() [ 704.701333] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 704.701333] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] return self._exit_event.wait() [ 704.701333] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 704.701333] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] current.throw(*self._exc) [ 704.701841] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 704.701841] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] result = function(*args, **kwargs) [ 704.701841] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 704.701841] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] return func(*args, **kwargs) [ 704.701841] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 704.701841] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] raise e [ 704.701841] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 704.701841] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] nwinfo = self.network_api.allocate_for_instance( [ 704.701841] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 704.701841] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] created_port_ids = self._update_ports_for_instance( [ 704.701841] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 704.701841] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] with excutils.save_and_reraise_exception(): [ 704.701841] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 704.702347] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] self.force_reraise() [ 704.702347] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 704.702347] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] raise self.value [ 704.702347] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 704.702347] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] updated_port = self._update_port( [ 704.702347] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 704.702347] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] _ensure_no_port_binding_failure(port) [ 704.702347] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 704.702347] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] raise exception.PortBindingFailed(port_id=port['id']) [ 704.702347] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] nova.exception.PortBindingFailed: Binding failed for port a1802ba1-ac36-4a83-a474-7c6e6253ca38, please check neutron logs for more information. [ 704.702347] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] [ 704.702347] env[63021]: INFO nova.compute.manager [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Terminating instance [ 704.705713] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Acquiring lock "refresh_cache-aef64522-2dbb-4bc5-84b2-a0bf8a108574" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.705843] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Acquired lock "refresh_cache-aef64522-2dbb-4bc5-84b2-a0bf8a108574" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.706019] env[63021]: DEBUG nova.network.neutron [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 704.757666] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-940dc987-4f7a-49fd-8602-43ab8541d6f6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.764511] env[63021]: DEBUG nova.compute.manager [req-b21f1d74-52bc-435f-9b7d-7e5c6f367a50 req-66d2d9ef-9275-467f-a9f3-c6d9b7855d6d service nova] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Received event network-vif-deleted-a1802ba1-ac36-4a83-a474-7c6e6253ca38 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 704.772769] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b114780a-bf71-4f39-b56c-e870e9b80986 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.804631] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1828fe8-a8ea-4002-a26f-c0d1260a5be1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.811945] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1bf082f-8cb0-4555-bc60-2638fdd5e988 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.829942] env[63021]: DEBUG nova.compute.provider_tree [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 705.238043] env[63021]: DEBUG nova.network.neutron [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 705.333524] env[63021]: DEBUG nova.compute.manager [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 705.335772] env[63021]: DEBUG nova.scheduler.client.report [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 705.365962] env[63021]: DEBUG nova.virt.hardware [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 705.365962] env[63021]: DEBUG nova.virt.hardware [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 705.365962] env[63021]: DEBUG nova.virt.hardware [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 705.366222] env[63021]: DEBUG nova.virt.hardware [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 705.366222] env[63021]: DEBUG nova.virt.hardware [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 705.366222] env[63021]: DEBUG nova.virt.hardware [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 705.366222] env[63021]: DEBUG nova.virt.hardware [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 705.366222] env[63021]: DEBUG nova.virt.hardware [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 705.366390] env[63021]: DEBUG nova.virt.hardware [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 705.366508] env[63021]: DEBUG nova.virt.hardware [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 705.366656] env[63021]: DEBUG nova.virt.hardware [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 705.368157] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c47060e-37f3-4769-aa08-d818642b216c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.377949] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b65f4968-5643-459d-9292-b0e6fe7844d9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.397503] env[63021]: DEBUG nova.network.neutron [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.722186] env[63021]: ERROR nova.compute.manager [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ea5445cf-f96e-48c9-8ee4-e1e62207c32e, please check neutron logs for more information. [ 705.722186] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 705.722186] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 705.722186] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 705.722186] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 705.722186] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 705.722186] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 705.722186] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 705.722186] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.722186] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 705.722186] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.722186] env[63021]: ERROR nova.compute.manager raise self.value [ 705.722186] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 705.722186] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 705.722186] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.722186] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 705.722901] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.722901] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 705.722901] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ea5445cf-f96e-48c9-8ee4-e1e62207c32e, please check neutron logs for more information. [ 705.722901] env[63021]: ERROR nova.compute.manager [ 705.722901] env[63021]: Traceback (most recent call last): [ 705.722901] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 705.722901] env[63021]: listener.cb(fileno) [ 705.722901] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 705.722901] env[63021]: result = function(*args, **kwargs) [ 705.722901] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 705.722901] env[63021]: return func(*args, **kwargs) [ 705.722901] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 705.722901] env[63021]: raise e [ 705.722901] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 705.722901] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 705.722901] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 705.722901] env[63021]: created_port_ids = self._update_ports_for_instance( [ 705.722901] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 705.722901] env[63021]: with excutils.save_and_reraise_exception(): [ 705.722901] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.722901] env[63021]: self.force_reraise() [ 705.722901] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.722901] env[63021]: raise self.value [ 705.722901] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 705.722901] env[63021]: updated_port = self._update_port( [ 705.722901] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.722901] env[63021]: _ensure_no_port_binding_failure(port) [ 705.722901] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.722901] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 705.724060] env[63021]: nova.exception.PortBindingFailed: Binding failed for port ea5445cf-f96e-48c9-8ee4-e1e62207c32e, please check neutron logs for more information. [ 705.724060] env[63021]: Removing descriptor: 18 [ 705.724060] env[63021]: ERROR nova.compute.manager [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ea5445cf-f96e-48c9-8ee4-e1e62207c32e, please check neutron logs for more information. [ 705.724060] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Traceback (most recent call last): [ 705.724060] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 705.724060] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] yield resources [ 705.724060] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 705.724060] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] self.driver.spawn(context, instance, image_meta, [ 705.724060] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 705.724060] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] self._vmops.spawn(context, instance, image_meta, injected_files, [ 705.724060] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 705.724060] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] vm_ref = self.build_virtual_machine(instance, [ 705.724451] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 705.724451] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] vif_infos = vmwarevif.get_vif_info(self._session, [ 705.724451] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 705.724451] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] for vif in network_info: [ 705.724451] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 705.724451] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] return self._sync_wrapper(fn, *args, **kwargs) [ 705.724451] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 705.724451] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] self.wait() [ 705.724451] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 705.724451] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] self[:] = self._gt.wait() [ 705.724451] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 705.724451] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] return self._exit_event.wait() [ 705.724451] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 705.724777] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] result = hub.switch() [ 705.724777] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 705.724777] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] return self.greenlet.switch() [ 705.724777] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 705.724777] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] result = function(*args, **kwargs) [ 705.724777] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 705.724777] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] return func(*args, **kwargs) [ 705.724777] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 705.724777] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] raise e [ 705.724777] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 705.724777] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] nwinfo = self.network_api.allocate_for_instance( [ 705.724777] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 705.724777] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] created_port_ids = self._update_ports_for_instance( [ 705.725116] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 705.725116] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] with excutils.save_and_reraise_exception(): [ 705.725116] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.725116] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] self.force_reraise() [ 705.725116] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.725116] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] raise self.value [ 705.725116] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 705.725116] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] updated_port = self._update_port( [ 705.725116] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.725116] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] _ensure_no_port_binding_failure(port) [ 705.725116] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.725116] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] raise exception.PortBindingFailed(port_id=port['id']) [ 705.725711] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] nova.exception.PortBindingFailed: Binding failed for port ea5445cf-f96e-48c9-8ee4-e1e62207c32e, please check neutron logs for more information. [ 705.725711] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] [ 705.725711] env[63021]: INFO nova.compute.manager [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Terminating instance [ 705.726835] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Acquiring lock "refresh_cache-c6ed1f3b-e80e-4545-8338-8e3aead5ba16" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.726835] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Acquired lock "refresh_cache-c6ed1f3b-e80e-4545-8338-8e3aead5ba16" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.726835] env[63021]: DEBUG nova.network.neutron [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 705.845921] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.545s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.845921] env[63021]: DEBUG nova.compute.manager [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 705.846484] env[63021]: DEBUG oslo_concurrency.lockutils [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.252s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.846484] env[63021]: DEBUG nova.objects.instance [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Lazy-loading 'resources' on Instance uuid 77fd1199-33b5-4040-a0ba-dd881da1224c {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 705.901886] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Releasing lock "refresh_cache-aef64522-2dbb-4bc5-84b2-a0bf8a108574" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.902568] env[63021]: DEBUG nova.compute.manager [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 705.902894] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-35d4eba1-54bc-46df-a648-3219828ed885 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.914252] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64a99429-f02f-4fd9-974f-9f6d67f1be1b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.941902] env[63021]: WARNING nova.virt.vmwareapi.driver [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance aef64522-2dbb-4bc5-84b2-a0bf8a108574 could not be found. [ 705.942178] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 705.942487] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-00a16087-823a-4f8b-9661-03588c66531b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.951426] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed4f4c5f-5f4a-4aef-9634-49dfc10ce9dd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.977684] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance aef64522-2dbb-4bc5-84b2-a0bf8a108574 could not be found. [ 705.977684] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 705.977684] env[63021]: INFO nova.compute.manager [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Took 0.07 seconds to destroy the instance on the hypervisor. [ 705.977684] env[63021]: DEBUG oslo.service.loopingcall [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 705.977684] env[63021]: DEBUG nova.compute.manager [-] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 705.977684] env[63021]: DEBUG nova.network.neutron [-] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 705.994752] env[63021]: DEBUG nova.network.neutron [-] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.265968] env[63021]: DEBUG nova.network.neutron [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.348132] env[63021]: DEBUG nova.compute.utils [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 706.349554] env[63021]: DEBUG nova.compute.manager [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 706.349715] env[63021]: DEBUG nova.network.neutron [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 706.357642] env[63021]: DEBUG nova.network.neutron [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.412137] env[63021]: DEBUG nova.policy [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '54b1c1b87ca84a20bbb43f8d47a54448', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b55ea247bf934099970e8d1da118abc6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 706.498440] env[63021]: DEBUG nova.network.neutron [-] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.790553] env[63021]: DEBUG nova.compute.manager [req-b5eb7d0d-649a-441c-82e8-97ae9ef41cc5 req-e47fa93b-8414-4b24-919f-bafbae69d77f service nova] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Received event network-changed-ea5445cf-f96e-48c9-8ee4-e1e62207c32e {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 706.790881] env[63021]: DEBUG nova.compute.manager [req-b5eb7d0d-649a-441c-82e8-97ae9ef41cc5 req-e47fa93b-8414-4b24-919f-bafbae69d77f service nova] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Refreshing instance network info cache due to event network-changed-ea5445cf-f96e-48c9-8ee4-e1e62207c32e. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 706.790967] env[63021]: DEBUG oslo_concurrency.lockutils [req-b5eb7d0d-649a-441c-82e8-97ae9ef41cc5 req-e47fa93b-8414-4b24-919f-bafbae69d77f service nova] Acquiring lock "refresh_cache-c6ed1f3b-e80e-4545-8338-8e3aead5ba16" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.797854] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b4f8c95-125f-421f-9d6c-0ad14d83cdc7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.806348] env[63021]: DEBUG nova.network.neutron [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Successfully created port: 5b7f5119-dade-4047-86cc-6cffab3760ad {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 706.808972] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34d16b8d-f496-4b0c-8aea-bf2b61366412 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.842841] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b978749-34cf-4846-b6da-f9272bb61ed1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.850229] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2acefd1-c078-4dd4-b3fb-c0459322aa34 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.855243] env[63021]: DEBUG nova.compute.manager [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 706.871709] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Releasing lock "refresh_cache-c6ed1f3b-e80e-4545-8338-8e3aead5ba16" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.871709] env[63021]: DEBUG nova.compute.manager [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 706.871709] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 706.871709] env[63021]: DEBUG nova.compute.provider_tree [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 706.871709] env[63021]: DEBUG oslo_concurrency.lockutils [req-b5eb7d0d-649a-441c-82e8-97ae9ef41cc5 req-e47fa93b-8414-4b24-919f-bafbae69d77f service nova] Acquired lock "refresh_cache-c6ed1f3b-e80e-4545-8338-8e3aead5ba16" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.872095] env[63021]: DEBUG nova.network.neutron [req-b5eb7d0d-649a-441c-82e8-97ae9ef41cc5 req-e47fa93b-8414-4b24-919f-bafbae69d77f service nova] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Refreshing network info cache for port ea5445cf-f96e-48c9-8ee4-e1e62207c32e {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 706.872095] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-19699746-334f-4f3f-aad3-4e7a1134182d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.880993] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a531993-be50-49e8-9ef9-6a72e0547a7c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.905184] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c6ed1f3b-e80e-4545-8338-8e3aead5ba16 could not be found. [ 706.905479] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 706.905676] env[63021]: INFO nova.compute.manager [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Took 0.04 seconds to destroy the instance on the hypervisor. [ 706.905976] env[63021]: DEBUG oslo.service.loopingcall [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 706.906250] env[63021]: DEBUG nova.compute.manager [-] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 706.906359] env[63021]: DEBUG nova.network.neutron [-] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 706.927736] env[63021]: DEBUG nova.network.neutron [-] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.001024] env[63021]: INFO nova.compute.manager [-] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Took 1.03 seconds to deallocate network for instance. [ 707.376690] env[63021]: DEBUG nova.scheduler.client.report [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 707.411160] env[63021]: DEBUG nova.network.neutron [req-b5eb7d0d-649a-441c-82e8-97ae9ef41cc5 req-e47fa93b-8414-4b24-919f-bafbae69d77f service nova] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.430226] env[63021]: DEBUG nova.network.neutron [-] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.526729] env[63021]: DEBUG nova.network.neutron [req-b5eb7d0d-649a-441c-82e8-97ae9ef41cc5 req-e47fa93b-8414-4b24-919f-bafbae69d77f service nova] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.568742] env[63021]: INFO nova.compute.manager [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Took 0.57 seconds to detach 1 volumes for instance. [ 707.570942] env[63021]: DEBUG nova.compute.claims [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 707.571408] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.722397] env[63021]: DEBUG nova.compute.manager [req-b1876093-e957-4ef5-8dcc-b807e6916479 req-b40614c0-c847-4807-a134-a94dcfe5ed20 service nova] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Received event network-changed-5b7f5119-dade-4047-86cc-6cffab3760ad {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 707.723186] env[63021]: DEBUG nova.compute.manager [req-b1876093-e957-4ef5-8dcc-b807e6916479 req-b40614c0-c847-4807-a134-a94dcfe5ed20 service nova] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Refreshing instance network info cache due to event network-changed-5b7f5119-dade-4047-86cc-6cffab3760ad. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 707.723186] env[63021]: DEBUG oslo_concurrency.lockutils [req-b1876093-e957-4ef5-8dcc-b807e6916479 req-b40614c0-c847-4807-a134-a94dcfe5ed20 service nova] Acquiring lock "refresh_cache-d36a9ed2-6849-43fc-ab88-77468bde337f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.723186] env[63021]: DEBUG oslo_concurrency.lockutils [req-b1876093-e957-4ef5-8dcc-b807e6916479 req-b40614c0-c847-4807-a134-a94dcfe5ed20 service nova] Acquired lock "refresh_cache-d36a9ed2-6849-43fc-ab88-77468bde337f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.723491] env[63021]: DEBUG nova.network.neutron [req-b1876093-e957-4ef5-8dcc-b807e6916479 req-b40614c0-c847-4807-a134-a94dcfe5ed20 service nova] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Refreshing network info cache for port 5b7f5119-dade-4047-86cc-6cffab3760ad {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 707.865896] env[63021]: DEBUG nova.compute.manager [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 707.886766] env[63021]: DEBUG oslo_concurrency.lockutils [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.040s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.894267] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.759s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.899816] env[63021]: INFO nova.compute.claims [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 707.911534] env[63021]: DEBUG nova.virt.hardware [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 707.911711] env[63021]: DEBUG nova.virt.hardware [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 707.911855] env[63021]: DEBUG nova.virt.hardware [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 707.912048] env[63021]: DEBUG nova.virt.hardware [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 707.912332] env[63021]: DEBUG nova.virt.hardware [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 707.912396] env[63021]: DEBUG nova.virt.hardware [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 707.912542] env[63021]: DEBUG nova.virt.hardware [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 707.912693] env[63021]: DEBUG nova.virt.hardware [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 707.912854] env[63021]: DEBUG nova.virt.hardware [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 707.913028] env[63021]: DEBUG nova.virt.hardware [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 707.913216] env[63021]: DEBUG nova.virt.hardware [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 707.914132] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e41cdf0-29fb-41b5-92b6-eb7d34bddc4a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.918618] env[63021]: INFO nova.scheduler.client.report [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Deleted allocations for instance 77fd1199-33b5-4040-a0ba-dd881da1224c [ 707.931315] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a6444f5-25c5-48ff-92c5-4771a6e20e1d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.938052] env[63021]: INFO nova.compute.manager [-] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Took 1.03 seconds to deallocate network for instance. [ 707.943604] env[63021]: DEBUG nova.compute.claims [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 707.943604] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.989542] env[63021]: ERROR nova.compute.manager [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5b7f5119-dade-4047-86cc-6cffab3760ad, please check neutron logs for more information. [ 707.989542] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 707.989542] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 707.989542] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 707.989542] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 707.989542] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 707.989542] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 707.989542] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 707.989542] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.989542] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 707.989542] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.989542] env[63021]: ERROR nova.compute.manager raise self.value [ 707.989542] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 707.989542] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 707.989542] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.989542] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 707.989992] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.989992] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 707.989992] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5b7f5119-dade-4047-86cc-6cffab3760ad, please check neutron logs for more information. [ 707.989992] env[63021]: ERROR nova.compute.manager [ 707.989992] env[63021]: Traceback (most recent call last): [ 707.989992] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 707.989992] env[63021]: listener.cb(fileno) [ 707.989992] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 707.989992] env[63021]: result = function(*args, **kwargs) [ 707.989992] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 707.989992] env[63021]: return func(*args, **kwargs) [ 707.989992] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 707.989992] env[63021]: raise e [ 707.989992] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 707.989992] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 707.989992] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 707.989992] env[63021]: created_port_ids = self._update_ports_for_instance( [ 707.989992] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 707.989992] env[63021]: with excutils.save_and_reraise_exception(): [ 707.989992] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.989992] env[63021]: self.force_reraise() [ 707.989992] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.989992] env[63021]: raise self.value [ 707.989992] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 707.989992] env[63021]: updated_port = self._update_port( [ 707.989992] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.989992] env[63021]: _ensure_no_port_binding_failure(port) [ 707.989992] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.989992] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 707.990712] env[63021]: nova.exception.PortBindingFailed: Binding failed for port 5b7f5119-dade-4047-86cc-6cffab3760ad, please check neutron logs for more information. [ 707.990712] env[63021]: Removing descriptor: 17 [ 707.990994] env[63021]: ERROR nova.compute.manager [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5b7f5119-dade-4047-86cc-6cffab3760ad, please check neutron logs for more information. [ 707.990994] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Traceback (most recent call last): [ 707.990994] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 707.990994] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] yield resources [ 707.990994] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 707.990994] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] self.driver.spawn(context, instance, image_meta, [ 707.990994] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 707.990994] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 707.990994] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 707.990994] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] vm_ref = self.build_virtual_machine(instance, [ 707.990994] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 707.991309] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] vif_infos = vmwarevif.get_vif_info(self._session, [ 707.991309] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 707.991309] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] for vif in network_info: [ 707.991309] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 707.991309] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] return self._sync_wrapper(fn, *args, **kwargs) [ 707.991309] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 707.991309] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] self.wait() [ 707.991309] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 707.991309] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] self[:] = self._gt.wait() [ 707.991309] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 707.991309] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] return self._exit_event.wait() [ 707.991309] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 707.991309] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] result = hub.switch() [ 707.991648] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 707.991648] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] return self.greenlet.switch() [ 707.991648] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 707.991648] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] result = function(*args, **kwargs) [ 707.991648] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 707.991648] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] return func(*args, **kwargs) [ 707.991648] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 707.991648] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] raise e [ 707.991648] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 707.991648] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] nwinfo = self.network_api.allocate_for_instance( [ 707.991648] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 707.991648] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] created_port_ids = self._update_ports_for_instance( [ 707.991648] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 707.991979] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] with excutils.save_and_reraise_exception(): [ 707.991979] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.991979] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] self.force_reraise() [ 707.991979] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.991979] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] raise self.value [ 707.991979] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 707.991979] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] updated_port = self._update_port( [ 707.991979] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.991979] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] _ensure_no_port_binding_failure(port) [ 707.991979] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.991979] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] raise exception.PortBindingFailed(port_id=port['id']) [ 707.991979] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] nova.exception.PortBindingFailed: Binding failed for port 5b7f5119-dade-4047-86cc-6cffab3760ad, please check neutron logs for more information. [ 707.991979] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] [ 707.992369] env[63021]: INFO nova.compute.manager [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Terminating instance [ 707.993387] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Acquiring lock "refresh_cache-d36a9ed2-6849-43fc-ab88-77468bde337f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.031655] env[63021]: DEBUG oslo_concurrency.lockutils [req-b5eb7d0d-649a-441c-82e8-97ae9ef41cc5 req-e47fa93b-8414-4b24-919f-bafbae69d77f service nova] Releasing lock "refresh_cache-c6ed1f3b-e80e-4545-8338-8e3aead5ba16" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.032581] env[63021]: DEBUG nova.compute.manager [req-b5eb7d0d-649a-441c-82e8-97ae9ef41cc5 req-e47fa93b-8414-4b24-919f-bafbae69d77f service nova] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Received event network-vif-deleted-ea5445cf-f96e-48c9-8ee4-e1e62207c32e {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 708.249768] env[63021]: DEBUG nova.network.neutron [req-b1876093-e957-4ef5-8dcc-b807e6916479 req-b40614c0-c847-4807-a134-a94dcfe5ed20 service nova] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 708.339493] env[63021]: DEBUG nova.network.neutron [req-b1876093-e957-4ef5-8dcc-b807e6916479 req-b40614c0-c847-4807-a134-a94dcfe5ed20 service nova] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.433933] env[63021]: DEBUG oslo_concurrency.lockutils [None req-38667f30-0428-4556-b8da-edfed9b5e481 tempest-ServersAaction247Test-2047804218 tempest-ServersAaction247Test-2047804218-project-member] Lock "77fd1199-33b5-4040-a0ba-dd881da1224c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.520s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.840863] env[63021]: DEBUG oslo_concurrency.lockutils [req-b1876093-e957-4ef5-8dcc-b807e6916479 req-b40614c0-c847-4807-a134-a94dcfe5ed20 service nova] Releasing lock "refresh_cache-d36a9ed2-6849-43fc-ab88-77468bde337f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.841286] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Acquired lock "refresh_cache-d36a9ed2-6849-43fc-ab88-77468bde337f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.841467] env[63021]: DEBUG nova.network.neutron [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 709.194553] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-981124e5-0d17-409f-b43f-1cfd1c8fe4a6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.202538] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-349b3422-96b4-4f0d-a7ac-cd57f7161be9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.235854] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90b2fc95-8b8a-4d0e-a02e-0880a17683a3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.243679] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d6302e5-cd68-432c-abff-6aa6bc41a0a6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.257233] env[63021]: DEBUG nova.compute.provider_tree [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 709.365320] env[63021]: DEBUG nova.network.neutron [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 709.460103] env[63021]: DEBUG nova.network.neutron [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.747677] env[63021]: DEBUG nova.compute.manager [req-99498913-2b7f-4ed5-9b6f-ad2f9aacc6a9 req-06666298-25b0-4537-bf97-d38b88060dd9 service nova] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Received event network-vif-deleted-5b7f5119-dade-4047-86cc-6cffab3760ad {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 709.759582] env[63021]: DEBUG nova.scheduler.client.report [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 709.968701] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Releasing lock "refresh_cache-d36a9ed2-6849-43fc-ab88-77468bde337f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.969174] env[63021]: DEBUG nova.compute.manager [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 709.969368] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 709.969710] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fdeccd68-49c1-4caf-a6fa-afb2bec37d51 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.980277] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e58e8f5-a125-4498-8bff-f917166bf92a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.009237] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d36a9ed2-6849-43fc-ab88-77468bde337f could not be found. [ 710.010006] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 710.010248] env[63021]: INFO nova.compute.manager [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 710.010563] env[63021]: DEBUG oslo.service.loopingcall [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 710.010792] env[63021]: DEBUG nova.compute.manager [-] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 710.010886] env[63021]: DEBUG nova.network.neutron [-] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 710.027378] env[63021]: DEBUG nova.network.neutron [-] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 710.267139] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.373s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.267667] env[63021]: DEBUG nova.compute.manager [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 710.270642] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.266s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.279738] env[63021]: INFO nova.compute.claims [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 710.532460] env[63021]: DEBUG nova.network.neutron [-] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.790232] env[63021]: DEBUG nova.compute.utils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 710.790232] env[63021]: DEBUG nova.compute.manager [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 710.790232] env[63021]: DEBUG nova.network.neutron [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 710.843888] env[63021]: DEBUG nova.policy [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '124196aaeeb6499dbfdee5d664e4421d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bb745825ae894ffc9ee752a8fbe18279', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 711.035518] env[63021]: INFO nova.compute.manager [-] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Took 1.02 seconds to deallocate network for instance. [ 711.038397] env[63021]: DEBUG nova.compute.claims [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 711.038627] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.290419] env[63021]: DEBUG nova.compute.manager [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 711.326846] env[63021]: DEBUG nova.network.neutron [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Successfully created port: 53c3f4ae-ad6b-431d-8d8c-50e340518f90 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 711.694644] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-801d058d-6891-4e2a-a6be-09dcfb9238d6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.702946] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-495f6826-aa49-4c8a-a36f-f37e065fd54c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.739296] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3798c474-be64-4198-b3d7-4093231ade8b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.746707] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35876282-4be5-4ece-8d83-a7173eda3ab0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.760438] env[63021]: DEBUG nova.compute.provider_tree [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 712.263573] env[63021]: DEBUG nova.scheduler.client.report [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 712.312487] env[63021]: DEBUG nova.compute.manager [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 712.339773] env[63021]: DEBUG nova.virt.hardware [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 712.339991] env[63021]: DEBUG nova.virt.hardware [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 712.340103] env[63021]: DEBUG nova.virt.hardware [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 712.340285] env[63021]: DEBUG nova.virt.hardware [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 712.340426] env[63021]: DEBUG nova.virt.hardware [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 712.340569] env[63021]: DEBUG nova.virt.hardware [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 712.340775] env[63021]: DEBUG nova.virt.hardware [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 712.340926] env[63021]: DEBUG nova.virt.hardware [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 712.341100] env[63021]: DEBUG nova.virt.hardware [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 712.341257] env[63021]: DEBUG nova.virt.hardware [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 712.341426] env[63021]: DEBUG nova.virt.hardware [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 712.342358] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e6d4c0c-b332-42d4-82e4-97c1f67edd37 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.350544] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a70ef914-55d4-4172-9dbf-5ff02db8c1a2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.432770] env[63021]: DEBUG nova.compute.manager [req-05f91284-b5a0-4103-a819-383290442fa0 req-ba296dfc-27f0-4be4-ace2-5210ed638ea9 service nova] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Received event network-changed-53c3f4ae-ad6b-431d-8d8c-50e340518f90 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 712.432944] env[63021]: DEBUG nova.compute.manager [req-05f91284-b5a0-4103-a819-383290442fa0 req-ba296dfc-27f0-4be4-ace2-5210ed638ea9 service nova] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Refreshing instance network info cache due to event network-changed-53c3f4ae-ad6b-431d-8d8c-50e340518f90. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 712.433206] env[63021]: DEBUG oslo_concurrency.lockutils [req-05f91284-b5a0-4103-a819-383290442fa0 req-ba296dfc-27f0-4be4-ace2-5210ed638ea9 service nova] Acquiring lock "refresh_cache-476e1e73-7050-4e53-bff8-d5d9f0982f1a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.433414] env[63021]: DEBUG oslo_concurrency.lockutils [req-05f91284-b5a0-4103-a819-383290442fa0 req-ba296dfc-27f0-4be4-ace2-5210ed638ea9 service nova] Acquired lock "refresh_cache-476e1e73-7050-4e53-bff8-d5d9f0982f1a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.433555] env[63021]: DEBUG nova.network.neutron [req-05f91284-b5a0-4103-a819-383290442fa0 req-ba296dfc-27f0-4be4-ace2-5210ed638ea9 service nova] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Refreshing network info cache for port 53c3f4ae-ad6b-431d-8d8c-50e340518f90 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 712.463340] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Acquiring lock "bef5c7cb-be2f-4be7-b72c-4a066d473dd0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.463635] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Lock "bef5c7cb-be2f-4be7-b72c-4a066d473dd0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.652717] env[63021]: ERROR nova.compute.manager [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 53c3f4ae-ad6b-431d-8d8c-50e340518f90, please check neutron logs for more information. [ 712.652717] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 712.652717] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.652717] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 712.652717] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 712.652717] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 712.652717] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 712.652717] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 712.652717] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.652717] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 712.652717] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.652717] env[63021]: ERROR nova.compute.manager raise self.value [ 712.652717] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 712.652717] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 712.652717] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.652717] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 712.653349] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.653349] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 712.653349] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 53c3f4ae-ad6b-431d-8d8c-50e340518f90, please check neutron logs for more information. [ 712.653349] env[63021]: ERROR nova.compute.manager [ 712.653349] env[63021]: Traceback (most recent call last): [ 712.653349] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 712.653349] env[63021]: listener.cb(fileno) [ 712.653349] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.653349] env[63021]: result = function(*args, **kwargs) [ 712.653349] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 712.653349] env[63021]: return func(*args, **kwargs) [ 712.653349] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 712.653349] env[63021]: raise e [ 712.653349] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.653349] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 712.653349] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 712.653349] env[63021]: created_port_ids = self._update_ports_for_instance( [ 712.653349] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 712.653349] env[63021]: with excutils.save_and_reraise_exception(): [ 712.653349] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.653349] env[63021]: self.force_reraise() [ 712.653349] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.653349] env[63021]: raise self.value [ 712.653349] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 712.653349] env[63021]: updated_port = self._update_port( [ 712.653349] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.653349] env[63021]: _ensure_no_port_binding_failure(port) [ 712.653349] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.653349] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 712.654848] env[63021]: nova.exception.PortBindingFailed: Binding failed for port 53c3f4ae-ad6b-431d-8d8c-50e340518f90, please check neutron logs for more information. [ 712.654848] env[63021]: Removing descriptor: 17 [ 712.654848] env[63021]: ERROR nova.compute.manager [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 53c3f4ae-ad6b-431d-8d8c-50e340518f90, please check neutron logs for more information. [ 712.654848] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Traceback (most recent call last): [ 712.654848] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 712.654848] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] yield resources [ 712.654848] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 712.654848] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] self.driver.spawn(context, instance, image_meta, [ 712.654848] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 712.654848] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 712.654848] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 712.654848] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] vm_ref = self.build_virtual_machine(instance, [ 712.655220] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 712.655220] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] vif_infos = vmwarevif.get_vif_info(self._session, [ 712.655220] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 712.655220] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] for vif in network_info: [ 712.655220] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 712.655220] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] return self._sync_wrapper(fn, *args, **kwargs) [ 712.655220] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 712.655220] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] self.wait() [ 712.655220] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 712.655220] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] self[:] = self._gt.wait() [ 712.655220] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 712.655220] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] return self._exit_event.wait() [ 712.655220] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 712.655590] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] result = hub.switch() [ 712.655590] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 712.655590] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] return self.greenlet.switch() [ 712.655590] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.655590] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] result = function(*args, **kwargs) [ 712.655590] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 712.655590] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] return func(*args, **kwargs) [ 712.655590] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 712.655590] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] raise e [ 712.655590] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.655590] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] nwinfo = self.network_api.allocate_for_instance( [ 712.655590] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 712.655590] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] created_port_ids = self._update_ports_for_instance( [ 712.655978] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 712.655978] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] with excutils.save_and_reraise_exception(): [ 712.655978] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.655978] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] self.force_reraise() [ 712.655978] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.655978] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] raise self.value [ 712.655978] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 712.655978] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] updated_port = self._update_port( [ 712.655978] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.655978] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] _ensure_no_port_binding_failure(port) [ 712.655978] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.655978] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] raise exception.PortBindingFailed(port_id=port['id']) [ 712.656329] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] nova.exception.PortBindingFailed: Binding failed for port 53c3f4ae-ad6b-431d-8d8c-50e340518f90, please check neutron logs for more information. [ 712.656329] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] [ 712.656329] env[63021]: INFO nova.compute.manager [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Terminating instance [ 712.656329] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquiring lock "refresh_cache-476e1e73-7050-4e53-bff8-d5d9f0982f1a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.768837] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.498s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.769415] env[63021]: DEBUG nova.compute.manager [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 712.773403] env[63021]: DEBUG oslo_concurrency.lockutils [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.876s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.775288] env[63021]: INFO nova.compute.claims [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 712.966994] env[63021]: DEBUG nova.network.neutron [req-05f91284-b5a0-4103-a819-383290442fa0 req-ba296dfc-27f0-4be4-ace2-5210ed638ea9 service nova] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 713.077511] env[63021]: DEBUG nova.network.neutron [req-05f91284-b5a0-4103-a819-383290442fa0 req-ba296dfc-27f0-4be4-ace2-5210ed638ea9 service nova] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.275189] env[63021]: DEBUG nova.compute.utils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 713.276574] env[63021]: DEBUG nova.compute.manager [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 713.276743] env[63021]: DEBUG nova.network.neutron [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 713.316886] env[63021]: DEBUG nova.policy [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '124196aaeeb6499dbfdee5d664e4421d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bb745825ae894ffc9ee752a8fbe18279', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 713.580511] env[63021]: DEBUG oslo_concurrency.lockutils [req-05f91284-b5a0-4103-a819-383290442fa0 req-ba296dfc-27f0-4be4-ace2-5210ed638ea9 service nova] Releasing lock "refresh_cache-476e1e73-7050-4e53-bff8-d5d9f0982f1a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.580838] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquired lock "refresh_cache-476e1e73-7050-4e53-bff8-d5d9f0982f1a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.581037] env[63021]: DEBUG nova.network.neutron [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 713.593960] env[63021]: DEBUG nova.network.neutron [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Successfully created port: 624e089c-f559-40e6-b38b-8abf34ede8a9 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 713.780179] env[63021]: DEBUG nova.compute.manager [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 714.106082] env[63021]: DEBUG nova.network.neutron [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.156155] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22063780-6d5a-4ce1-a1fa-a5980a5a1170 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.169384] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c8eca0e-74ac-48cf-858e-20f521b828be {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.200991] env[63021]: DEBUG nova.network.neutron [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.202780] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-202a18b0-5153-4d84-82a1-793709780125 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.212267] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-088531e3-d5c7-42b5-8dce-1f79af01c878 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.229141] env[63021]: DEBUG nova.compute.provider_tree [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 714.459900] env[63021]: DEBUG nova.compute.manager [req-3e813381-b5d7-4c4e-bc62-59753f965246 req-b27deebd-c3f9-49f3-8201-60077ce21255 service nova] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Received event network-vif-deleted-53c3f4ae-ad6b-431d-8d8c-50e340518f90 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 714.697281] env[63021]: ERROR nova.compute.manager [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 624e089c-f559-40e6-b38b-8abf34ede8a9, please check neutron logs for more information. [ 714.697281] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 714.697281] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 714.697281] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 714.697281] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 714.697281] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 714.697281] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 714.697281] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 714.697281] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 714.697281] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 714.697281] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 714.697281] env[63021]: ERROR nova.compute.manager raise self.value [ 714.697281] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 714.697281] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 714.697281] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 714.697281] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 714.697720] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 714.697720] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 714.697720] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 624e089c-f559-40e6-b38b-8abf34ede8a9, please check neutron logs for more information. [ 714.697720] env[63021]: ERROR nova.compute.manager [ 714.697720] env[63021]: Traceback (most recent call last): [ 714.697720] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 714.697720] env[63021]: listener.cb(fileno) [ 714.697720] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 714.697720] env[63021]: result = function(*args, **kwargs) [ 714.697720] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 714.697720] env[63021]: return func(*args, **kwargs) [ 714.697720] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 714.697720] env[63021]: raise e [ 714.697720] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 714.697720] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 714.697720] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 714.697720] env[63021]: created_port_ids = self._update_ports_for_instance( [ 714.697720] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 714.697720] env[63021]: with excutils.save_and_reraise_exception(): [ 714.697720] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 714.697720] env[63021]: self.force_reraise() [ 714.697720] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 714.697720] env[63021]: raise self.value [ 714.697720] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 714.697720] env[63021]: updated_port = self._update_port( [ 714.697720] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 714.697720] env[63021]: _ensure_no_port_binding_failure(port) [ 714.697720] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 714.697720] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 714.698551] env[63021]: nova.exception.PortBindingFailed: Binding failed for port 624e089c-f559-40e6-b38b-8abf34ede8a9, please check neutron logs for more information. [ 714.698551] env[63021]: Removing descriptor: 17 [ 714.707279] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Releasing lock "refresh_cache-476e1e73-7050-4e53-bff8-d5d9f0982f1a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.707710] env[63021]: DEBUG nova.compute.manager [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 714.707913] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 714.708225] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3e7daaa4-4951-4b78-94a3-de609c0c4726 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.717338] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8f77af3-130d-4bac-8711-1a25eae1a461 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.731722] env[63021]: DEBUG nova.scheduler.client.report [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 714.739594] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 476e1e73-7050-4e53-bff8-d5d9f0982f1a could not be found. [ 714.739784] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 714.739953] env[63021]: INFO nova.compute.manager [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Took 0.03 seconds to destroy the instance on the hypervisor. [ 714.740210] env[63021]: DEBUG oslo.service.loopingcall [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 714.740416] env[63021]: DEBUG nova.compute.manager [-] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 714.740502] env[63021]: DEBUG nova.network.neutron [-] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 714.754472] env[63021]: DEBUG nova.network.neutron [-] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.792491] env[63021]: DEBUG nova.compute.manager [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 714.817540] env[63021]: DEBUG nova.virt.hardware [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 714.817744] env[63021]: DEBUG nova.virt.hardware [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 714.817899] env[63021]: DEBUG nova.virt.hardware [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 714.818086] env[63021]: DEBUG nova.virt.hardware [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 714.818234] env[63021]: DEBUG nova.virt.hardware [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 714.818381] env[63021]: DEBUG nova.virt.hardware [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 714.818584] env[63021]: DEBUG nova.virt.hardware [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 714.818739] env[63021]: DEBUG nova.virt.hardware [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 714.818901] env[63021]: DEBUG nova.virt.hardware [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 714.819069] env[63021]: DEBUG nova.virt.hardware [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 714.819243] env[63021]: DEBUG nova.virt.hardware [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 714.820093] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15608141-8dc7-49cf-aecd-9e9efcfabbff {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.828525] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-220cf119-dac5-4476-9867-272895e5b1ab {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.843198] env[63021]: ERROR nova.compute.manager [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 624e089c-f559-40e6-b38b-8abf34ede8a9, please check neutron logs for more information. [ 714.843198] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Traceback (most recent call last): [ 714.843198] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 714.843198] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] yield resources [ 714.843198] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 714.843198] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] self.driver.spawn(context, instance, image_meta, [ 714.843198] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 714.843198] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 714.843198] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 714.843198] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] vm_ref = self.build_virtual_machine(instance, [ 714.843198] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 714.843615] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] vif_infos = vmwarevif.get_vif_info(self._session, [ 714.843615] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 714.843615] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] for vif in network_info: [ 714.843615] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 714.843615] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] return self._sync_wrapper(fn, *args, **kwargs) [ 714.843615] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 714.843615] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] self.wait() [ 714.843615] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 714.843615] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] self[:] = self._gt.wait() [ 714.843615] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 714.843615] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] return self._exit_event.wait() [ 714.843615] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 714.843615] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] current.throw(*self._exc) [ 714.844055] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 714.844055] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] result = function(*args, **kwargs) [ 714.844055] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 714.844055] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] return func(*args, **kwargs) [ 714.844055] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 714.844055] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] raise e [ 714.844055] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 714.844055] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] nwinfo = self.network_api.allocate_for_instance( [ 714.844055] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 714.844055] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] created_port_ids = self._update_ports_for_instance( [ 714.844055] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 714.844055] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] with excutils.save_and_reraise_exception(): [ 714.844055] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 714.844469] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] self.force_reraise() [ 714.844469] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 714.844469] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] raise self.value [ 714.844469] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 714.844469] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] updated_port = self._update_port( [ 714.844469] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 714.844469] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] _ensure_no_port_binding_failure(port) [ 714.844469] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 714.844469] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] raise exception.PortBindingFailed(port_id=port['id']) [ 714.844469] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] nova.exception.PortBindingFailed: Binding failed for port 624e089c-f559-40e6-b38b-8abf34ede8a9, please check neutron logs for more information. [ 714.844469] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] [ 714.844469] env[63021]: INFO nova.compute.manager [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Terminating instance [ 714.845373] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquiring lock "refresh_cache-28d70e1d-da2a-4c9c-b297-86d026ce42c3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.845532] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquired lock "refresh_cache-28d70e1d-da2a-4c9c-b297-86d026ce42c3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.845681] env[63021]: DEBUG nova.network.neutron [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 715.239234] env[63021]: DEBUG oslo_concurrency.lockutils [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.463s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.239234] env[63021]: DEBUG nova.compute.manager [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 715.240812] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.433s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.257793] env[63021]: DEBUG nova.network.neutron [-] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.362723] env[63021]: DEBUG nova.network.neutron [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 715.457388] env[63021]: DEBUG nova.network.neutron [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.745440] env[63021]: DEBUG nova.compute.utils [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 715.749757] env[63021]: DEBUG nova.compute.manager [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 715.749933] env[63021]: DEBUG nova.network.neutron [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 715.760129] env[63021]: INFO nova.compute.manager [-] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Took 1.02 seconds to deallocate network for instance. [ 715.763007] env[63021]: DEBUG nova.compute.claims [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 715.763258] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.805120] env[63021]: DEBUG nova.policy [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e3b386c5f3b547ba9fa6a59b64bfc8e2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '455565d02ead4c4bb04c55713521ff39', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 715.960142] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Releasing lock "refresh_cache-28d70e1d-da2a-4c9c-b297-86d026ce42c3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.960533] env[63021]: DEBUG nova.compute.manager [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 715.960667] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 715.960959] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8b9b8f8b-62e8-4e73-a1fb-be3419c087be {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.970423] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b8b718d-a742-4d0f-97e6-f0170a2ad8ef {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.994100] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 28d70e1d-da2a-4c9c-b297-86d026ce42c3 could not be found. [ 715.994360] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 715.994596] env[63021]: INFO nova.compute.manager [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Took 0.03 seconds to destroy the instance on the hypervisor. [ 715.994877] env[63021]: DEBUG oslo.service.loopingcall [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 715.997390] env[63021]: DEBUG nova.compute.manager [-] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 715.997498] env[63021]: DEBUG nova.network.neutron [-] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 716.013719] env[63021]: DEBUG nova.network.neutron [-] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 716.072299] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05b35e39-f9d9-4a01-9d64-d69c2181fe3d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.079633] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11eab981-9390-48e6-b720-f44b1f8fe665 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.113997] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f0e43a2-2611-4e5b-9fd7-7d5cbff1b41f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.121305] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af3d267b-185e-4664-8ebd-70fa3edc5e80 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.134371] env[63021]: DEBUG nova.compute.provider_tree [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 716.156757] env[63021]: DEBUG nova.network.neutron [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Successfully created port: 9c006af2-d842-42f6-9191-74570d1de950 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 716.250711] env[63021]: DEBUG nova.compute.manager [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 716.487317] env[63021]: DEBUG nova.compute.manager [req-a8ec1258-7d64-4da4-b75c-44239b056de4 req-34500cd4-31a5-4287-9240-585479084a81 service nova] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Received event network-changed-624e089c-f559-40e6-b38b-8abf34ede8a9 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 716.487317] env[63021]: DEBUG nova.compute.manager [req-a8ec1258-7d64-4da4-b75c-44239b056de4 req-34500cd4-31a5-4287-9240-585479084a81 service nova] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Refreshing instance network info cache due to event network-changed-624e089c-f559-40e6-b38b-8abf34ede8a9. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 716.487317] env[63021]: DEBUG oslo_concurrency.lockutils [req-a8ec1258-7d64-4da4-b75c-44239b056de4 req-34500cd4-31a5-4287-9240-585479084a81 service nova] Acquiring lock "refresh_cache-28d70e1d-da2a-4c9c-b297-86d026ce42c3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.487317] env[63021]: DEBUG oslo_concurrency.lockutils [req-a8ec1258-7d64-4da4-b75c-44239b056de4 req-34500cd4-31a5-4287-9240-585479084a81 service nova] Acquired lock "refresh_cache-28d70e1d-da2a-4c9c-b297-86d026ce42c3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.487317] env[63021]: DEBUG nova.network.neutron [req-a8ec1258-7d64-4da4-b75c-44239b056de4 req-34500cd4-31a5-4287-9240-585479084a81 service nova] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Refreshing network info cache for port 624e089c-f559-40e6-b38b-8abf34ede8a9 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 716.517922] env[63021]: DEBUG nova.network.neutron [-] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.639858] env[63021]: DEBUG nova.scheduler.client.report [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 717.006548] env[63021]: DEBUG nova.network.neutron [req-a8ec1258-7d64-4da4-b75c-44239b056de4 req-34500cd4-31a5-4287-9240-585479084a81 service nova] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 717.022256] env[63021]: INFO nova.compute.manager [-] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Took 1.02 seconds to deallocate network for instance. [ 717.024533] env[63021]: DEBUG nova.compute.claims [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 717.024708] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.087193] env[63021]: DEBUG nova.network.neutron [req-a8ec1258-7d64-4da4-b75c-44239b056de4 req-34500cd4-31a5-4287-9240-585479084a81 service nova] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.141851] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.901s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.144097] env[63021]: ERROR nova.compute.manager [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 670c7666-bae2-4f9f-b75c-9a31a67a5b44, please check neutron logs for more information. [ 717.144097] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Traceback (most recent call last): [ 717.144097] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 717.144097] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] self.driver.spawn(context, instance, image_meta, [ 717.144097] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 717.144097] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] self._vmops.spawn(context, instance, image_meta, injected_files, [ 717.144097] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 717.144097] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] vm_ref = self.build_virtual_machine(instance, [ 717.144097] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 717.144097] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] vif_infos = vmwarevif.get_vif_info(self._session, [ 717.144097] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 717.144703] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] for vif in network_info: [ 717.144703] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 717.144703] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] return self._sync_wrapper(fn, *args, **kwargs) [ 717.144703] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 717.144703] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] self.wait() [ 717.144703] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 717.144703] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] self[:] = self._gt.wait() [ 717.144703] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 717.144703] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] return self._exit_event.wait() [ 717.144703] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 717.144703] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] current.throw(*self._exc) [ 717.144703] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 717.144703] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] result = function(*args, **kwargs) [ 717.145326] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 717.145326] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] return func(*args, **kwargs) [ 717.145326] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 717.145326] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] raise e [ 717.145326] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 717.145326] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] nwinfo = self.network_api.allocate_for_instance( [ 717.145326] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 717.145326] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] created_port_ids = self._update_ports_for_instance( [ 717.145326] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 717.145326] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] with excutils.save_and_reraise_exception(): [ 717.145326] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.145326] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] self.force_reraise() [ 717.145326] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.146032] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] raise self.value [ 717.146032] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 717.146032] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] updated_port = self._update_port( [ 717.146032] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.146032] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] _ensure_no_port_binding_failure(port) [ 717.146032] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.146032] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] raise exception.PortBindingFailed(port_id=port['id']) [ 717.146032] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] nova.exception.PortBindingFailed: Binding failed for port 670c7666-bae2-4f9f-b75c-9a31a67a5b44, please check neutron logs for more information. [ 717.146032] env[63021]: ERROR nova.compute.manager [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] [ 717.146032] env[63021]: DEBUG nova.compute.utils [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Binding failed for port 670c7666-bae2-4f9f-b75c-9a31a67a5b44, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 717.146520] env[63021]: DEBUG oslo_concurrency.lockutils [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.849s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.146520] env[63021]: INFO nova.compute.claims [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 717.151024] env[63021]: DEBUG nova.compute.manager [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Build of instance 52b854c4-ccc8-4e38-8100-31ab06959939 was re-scheduled: Binding failed for port 670c7666-bae2-4f9f-b75c-9a31a67a5b44, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 717.151024] env[63021]: DEBUG nova.compute.manager [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 717.151024] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "refresh_cache-52b854c4-ccc8-4e38-8100-31ab06959939" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.151024] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquired lock "refresh_cache-52b854c4-ccc8-4e38-8100-31ab06959939" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.151231] env[63021]: DEBUG nova.network.neutron [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 717.259899] env[63021]: DEBUG nova.compute.manager [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 717.287030] env[63021]: DEBUG nova.virt.hardware [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 717.287030] env[63021]: DEBUG nova.virt.hardware [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 717.287030] env[63021]: DEBUG nova.virt.hardware [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 717.287417] env[63021]: DEBUG nova.virt.hardware [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 717.287417] env[63021]: DEBUG nova.virt.hardware [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 717.287417] env[63021]: DEBUG nova.virt.hardware [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 717.287417] env[63021]: DEBUG nova.virt.hardware [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 717.287417] env[63021]: DEBUG nova.virt.hardware [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 717.287566] env[63021]: DEBUG nova.virt.hardware [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 717.287566] env[63021]: DEBUG nova.virt.hardware [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 717.287566] env[63021]: DEBUG nova.virt.hardware [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 717.288531] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e48d616-754a-47ba-b9ed-ea38b5af7220 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.297538] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e97d3b5c-f157-4e04-be77-e92492adf3b9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.429330] env[63021]: ERROR nova.compute.manager [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9c006af2-d842-42f6-9191-74570d1de950, please check neutron logs for more information. [ 717.429330] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 717.429330] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 717.429330] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 717.429330] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 717.429330] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 717.429330] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 717.429330] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 717.429330] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.429330] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 717.429330] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.429330] env[63021]: ERROR nova.compute.manager raise self.value [ 717.429330] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 717.429330] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 717.429330] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.429330] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 717.430025] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.430025] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 717.430025] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9c006af2-d842-42f6-9191-74570d1de950, please check neutron logs for more information. [ 717.430025] env[63021]: ERROR nova.compute.manager [ 717.430025] env[63021]: Traceback (most recent call last): [ 717.430025] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 717.430025] env[63021]: listener.cb(fileno) [ 717.430025] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 717.430025] env[63021]: result = function(*args, **kwargs) [ 717.430025] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 717.430025] env[63021]: return func(*args, **kwargs) [ 717.430025] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 717.430025] env[63021]: raise e [ 717.430025] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 717.430025] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 717.430025] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 717.430025] env[63021]: created_port_ids = self._update_ports_for_instance( [ 717.430025] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 717.430025] env[63021]: with excutils.save_and_reraise_exception(): [ 717.430025] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.430025] env[63021]: self.force_reraise() [ 717.430025] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.430025] env[63021]: raise self.value [ 717.430025] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 717.430025] env[63021]: updated_port = self._update_port( [ 717.430025] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.430025] env[63021]: _ensure_no_port_binding_failure(port) [ 717.430025] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.430025] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 717.431186] env[63021]: nova.exception.PortBindingFailed: Binding failed for port 9c006af2-d842-42f6-9191-74570d1de950, please check neutron logs for more information. [ 717.431186] env[63021]: Removing descriptor: 17 [ 717.431186] env[63021]: ERROR nova.compute.manager [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9c006af2-d842-42f6-9191-74570d1de950, please check neutron logs for more information. [ 717.431186] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Traceback (most recent call last): [ 717.431186] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 717.431186] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] yield resources [ 717.431186] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 717.431186] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] self.driver.spawn(context, instance, image_meta, [ 717.431186] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 717.431186] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 717.431186] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 717.431186] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] vm_ref = self.build_virtual_machine(instance, [ 717.431682] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 717.431682] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] vif_infos = vmwarevif.get_vif_info(self._session, [ 717.431682] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 717.431682] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] for vif in network_info: [ 717.431682] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 717.431682] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] return self._sync_wrapper(fn, *args, **kwargs) [ 717.431682] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 717.431682] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] self.wait() [ 717.431682] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 717.431682] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] self[:] = self._gt.wait() [ 717.431682] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 717.431682] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] return self._exit_event.wait() [ 717.431682] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 717.432268] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] result = hub.switch() [ 717.432268] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 717.432268] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] return self.greenlet.switch() [ 717.432268] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 717.432268] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] result = function(*args, **kwargs) [ 717.432268] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 717.432268] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] return func(*args, **kwargs) [ 717.432268] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 717.432268] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] raise e [ 717.432268] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 717.432268] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] nwinfo = self.network_api.allocate_for_instance( [ 717.432268] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 717.432268] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] created_port_ids = self._update_ports_for_instance( [ 717.432795] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 717.432795] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] with excutils.save_and_reraise_exception(): [ 717.432795] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.432795] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] self.force_reraise() [ 717.432795] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.432795] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] raise self.value [ 717.432795] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 717.432795] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] updated_port = self._update_port( [ 717.432795] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.432795] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] _ensure_no_port_binding_failure(port) [ 717.432795] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.432795] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] raise exception.PortBindingFailed(port_id=port['id']) [ 717.433341] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] nova.exception.PortBindingFailed: Binding failed for port 9c006af2-d842-42f6-9191-74570d1de950, please check neutron logs for more information. [ 717.433341] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] [ 717.433341] env[63021]: INFO nova.compute.manager [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Terminating instance [ 717.433341] env[63021]: DEBUG oslo_concurrency.lockutils [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Acquiring lock "refresh_cache-03aec3ed-f338-41b5-a396-e3d693edbb8c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.433341] env[63021]: DEBUG oslo_concurrency.lockutils [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Acquired lock "refresh_cache-03aec3ed-f338-41b5-a396-e3d693edbb8c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.433341] env[63021]: DEBUG nova.network.neutron [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 717.588544] env[63021]: DEBUG oslo_concurrency.lockutils [req-a8ec1258-7d64-4da4-b75c-44239b056de4 req-34500cd4-31a5-4287-9240-585479084a81 service nova] Releasing lock "refresh_cache-28d70e1d-da2a-4c9c-b297-86d026ce42c3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.588863] env[63021]: DEBUG nova.compute.manager [req-a8ec1258-7d64-4da4-b75c-44239b056de4 req-34500cd4-31a5-4287-9240-585479084a81 service nova] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Received event network-vif-deleted-624e089c-f559-40e6-b38b-8abf34ede8a9 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 717.670114] env[63021]: DEBUG nova.network.neutron [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 717.762513] env[63021]: DEBUG nova.network.neutron [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.950616] env[63021]: DEBUG nova.network.neutron [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 718.019116] env[63021]: DEBUG nova.network.neutron [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.266387] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Releasing lock "refresh_cache-52b854c4-ccc8-4e38-8100-31ab06959939" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.266387] env[63021]: DEBUG nova.compute.manager [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 718.266574] env[63021]: DEBUG nova.compute.manager [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 718.266645] env[63021]: DEBUG nova.network.neutron [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 718.291291] env[63021]: DEBUG nova.network.neutron [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 718.430231] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-758eb3a1-0cbf-4c0a-b218-24bf4b30295d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.439021] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5f35532-8ceb-4533-bbd2-3aa74502b33c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.467882] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42b7ec24-376e-4711-a3ab-dc8535c6b89d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.476220] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05e76419-6d22-42b5-b90b-210c57085111 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.493313] env[63021]: DEBUG nova.compute.provider_tree [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 718.522160] env[63021]: DEBUG oslo_concurrency.lockutils [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Releasing lock "refresh_cache-03aec3ed-f338-41b5-a396-e3d693edbb8c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.522625] env[63021]: DEBUG nova.compute.manager [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 718.523261] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 718.523261] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-144205cf-9fb8-4c02-9217-3f5249486780 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.531961] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-730fc82c-a586-45c5-a5e7-b9febf6828ce {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.547769] env[63021]: DEBUG nova.compute.manager [req-0d95d954-f91d-448e-8db2-407c06368392 req-269102d0-7aaf-47ac-aec8-98f684dfd52b service nova] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Received event network-changed-9c006af2-d842-42f6-9191-74570d1de950 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 718.547862] env[63021]: DEBUG nova.compute.manager [req-0d95d954-f91d-448e-8db2-407c06368392 req-269102d0-7aaf-47ac-aec8-98f684dfd52b service nova] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Refreshing instance network info cache due to event network-changed-9c006af2-d842-42f6-9191-74570d1de950. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 718.548062] env[63021]: DEBUG oslo_concurrency.lockutils [req-0d95d954-f91d-448e-8db2-407c06368392 req-269102d0-7aaf-47ac-aec8-98f684dfd52b service nova] Acquiring lock "refresh_cache-03aec3ed-f338-41b5-a396-e3d693edbb8c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.548235] env[63021]: DEBUG oslo_concurrency.lockutils [req-0d95d954-f91d-448e-8db2-407c06368392 req-269102d0-7aaf-47ac-aec8-98f684dfd52b service nova] Acquired lock "refresh_cache-03aec3ed-f338-41b5-a396-e3d693edbb8c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.548368] env[63021]: DEBUG nova.network.neutron [req-0d95d954-f91d-448e-8db2-407c06368392 req-269102d0-7aaf-47ac-aec8-98f684dfd52b service nova] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Refreshing network info cache for port 9c006af2-d842-42f6-9191-74570d1de950 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 718.553443] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 03aec3ed-f338-41b5-a396-e3d693edbb8c could not be found. [ 718.553637] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 718.553804] env[63021]: INFO nova.compute.manager [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Took 0.03 seconds to destroy the instance on the hypervisor. [ 718.554121] env[63021]: DEBUG oslo.service.loopingcall [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 718.554507] env[63021]: DEBUG nova.compute.manager [-] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 718.554602] env[63021]: DEBUG nova.network.neutron [-] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 718.574869] env[63021]: DEBUG nova.network.neutron [-] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 718.793803] env[63021]: DEBUG nova.network.neutron [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.996334] env[63021]: DEBUG nova.scheduler.client.report [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 719.072303] env[63021]: DEBUG nova.network.neutron [req-0d95d954-f91d-448e-8db2-407c06368392 req-269102d0-7aaf-47ac-aec8-98f684dfd52b service nova] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.076787] env[63021]: DEBUG nova.network.neutron [-] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.188753] env[63021]: DEBUG nova.network.neutron [req-0d95d954-f91d-448e-8db2-407c06368392 req-269102d0-7aaf-47ac-aec8-98f684dfd52b service nova] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.297297] env[63021]: INFO nova.compute.manager [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 52b854c4-ccc8-4e38-8100-31ab06959939] Took 1.03 seconds to deallocate network for instance. [ 719.501138] env[63021]: DEBUG oslo_concurrency.lockutils [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.356s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.501690] env[63021]: DEBUG nova.compute.manager [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 719.504814] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.214s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.506609] env[63021]: INFO nova.compute.claims [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 719.579239] env[63021]: INFO nova.compute.manager [-] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Took 1.02 seconds to deallocate network for instance. [ 719.581582] env[63021]: DEBUG nova.compute.claims [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 719.581757] env[63021]: DEBUG oslo_concurrency.lockutils [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.691761] env[63021]: DEBUG oslo_concurrency.lockutils [req-0d95d954-f91d-448e-8db2-407c06368392 req-269102d0-7aaf-47ac-aec8-98f684dfd52b service nova] Releasing lock "refresh_cache-03aec3ed-f338-41b5-a396-e3d693edbb8c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.692059] env[63021]: DEBUG nova.compute.manager [req-0d95d954-f91d-448e-8db2-407c06368392 req-269102d0-7aaf-47ac-aec8-98f684dfd52b service nova] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Received event network-vif-deleted-9c006af2-d842-42f6-9191-74570d1de950 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 720.014065] env[63021]: DEBUG nova.compute.utils [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 720.015387] env[63021]: DEBUG nova.compute.manager [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 720.015564] env[63021]: DEBUG nova.network.neutron [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] [instance: 239d0087-462d-445a-a131-a23d4d15505f] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 720.053252] env[63021]: DEBUG nova.policy [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f6d65ec3ce574a4c9e6a9aa291017d90', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4abe19a1a8d44e65882f8d7f9674e670', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 720.330927] env[63021]: INFO nova.scheduler.client.report [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Deleted allocations for instance 52b854c4-ccc8-4e38-8100-31ab06959939 [ 720.373216] env[63021]: DEBUG nova.network.neutron [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Successfully created port: f20d29e3-a559-4952-9e91-3afed97b91b6 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 720.519955] env[63021]: DEBUG nova.compute.manager [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 720.798277] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28d5725d-c403-4dc8-aca2-bbbce3d0c79c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.806385] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20eba723-3196-4af5-913e-2ad8d9ccc37f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.837272] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf13cb96-af65-4174-b1db-bc462705605f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.842388] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4fb60ac9-dcd6-40ae-8b89-80d3348b775b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "52b854c4-ccc8-4e38-8100-31ab06959939" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.416s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.847288] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbc4b0ac-9ba7-492d-9eba-61f19104d41e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.861865] env[63021]: DEBUG nova.compute.provider_tree [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 721.273372] env[63021]: DEBUG nova.compute.manager [req-6c2a07c3-0c35-49a9-aa68-d87f30f49053 req-82d8b557-6014-4f41-bc00-790af42986b4 service nova] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Received event network-changed-f20d29e3-a559-4952-9e91-3afed97b91b6 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 721.273372] env[63021]: DEBUG nova.compute.manager [req-6c2a07c3-0c35-49a9-aa68-d87f30f49053 req-82d8b557-6014-4f41-bc00-790af42986b4 service nova] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Refreshing instance network info cache due to event network-changed-f20d29e3-a559-4952-9e91-3afed97b91b6. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 721.273509] env[63021]: DEBUG oslo_concurrency.lockutils [req-6c2a07c3-0c35-49a9-aa68-d87f30f49053 req-82d8b557-6014-4f41-bc00-790af42986b4 service nova] Acquiring lock "refresh_cache-239d0087-462d-445a-a131-a23d4d15505f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.273578] env[63021]: DEBUG oslo_concurrency.lockutils [req-6c2a07c3-0c35-49a9-aa68-d87f30f49053 req-82d8b557-6014-4f41-bc00-790af42986b4 service nova] Acquired lock "refresh_cache-239d0087-462d-445a-a131-a23d4d15505f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.273720] env[63021]: DEBUG nova.network.neutron [req-6c2a07c3-0c35-49a9-aa68-d87f30f49053 req-82d8b557-6014-4f41-bc00-790af42986b4 service nova] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Refreshing network info cache for port f20d29e3-a559-4952-9e91-3afed97b91b6 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 721.347375] env[63021]: DEBUG nova.compute.manager [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 721.364439] env[63021]: DEBUG nova.scheduler.client.report [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 721.448075] env[63021]: ERROR nova.compute.manager [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f20d29e3-a559-4952-9e91-3afed97b91b6, please check neutron logs for more information. [ 721.448075] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 721.448075] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 721.448075] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 721.448075] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 721.448075] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 721.448075] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 721.448075] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 721.448075] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.448075] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 721.448075] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.448075] env[63021]: ERROR nova.compute.manager raise self.value [ 721.448075] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 721.448075] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 721.448075] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.448075] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 721.449771] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.449771] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 721.449771] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f20d29e3-a559-4952-9e91-3afed97b91b6, please check neutron logs for more information. [ 721.449771] env[63021]: ERROR nova.compute.manager [ 721.449771] env[63021]: Traceback (most recent call last): [ 721.449771] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 721.449771] env[63021]: listener.cb(fileno) [ 721.449771] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 721.449771] env[63021]: result = function(*args, **kwargs) [ 721.449771] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 721.449771] env[63021]: return func(*args, **kwargs) [ 721.449771] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 721.449771] env[63021]: raise e [ 721.449771] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 721.449771] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 721.449771] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 721.449771] env[63021]: created_port_ids = self._update_ports_for_instance( [ 721.449771] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 721.449771] env[63021]: with excutils.save_and_reraise_exception(): [ 721.449771] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.449771] env[63021]: self.force_reraise() [ 721.449771] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.449771] env[63021]: raise self.value [ 721.449771] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 721.449771] env[63021]: updated_port = self._update_port( [ 721.449771] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.449771] env[63021]: _ensure_no_port_binding_failure(port) [ 721.449771] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.449771] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 721.452579] env[63021]: nova.exception.PortBindingFailed: Binding failed for port f20d29e3-a559-4952-9e91-3afed97b91b6, please check neutron logs for more information. [ 721.452579] env[63021]: Removing descriptor: 17 [ 721.532476] env[63021]: DEBUG nova.compute.manager [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 721.556213] env[63021]: DEBUG nova.virt.hardware [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 721.556450] env[63021]: DEBUG nova.virt.hardware [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 721.556603] env[63021]: DEBUG nova.virt.hardware [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 721.556778] env[63021]: DEBUG nova.virt.hardware [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 721.556918] env[63021]: DEBUG nova.virt.hardware [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 721.557070] env[63021]: DEBUG nova.virt.hardware [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 721.557443] env[63021]: DEBUG nova.virt.hardware [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 721.558076] env[63021]: DEBUG nova.virt.hardware [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 721.558076] env[63021]: DEBUG nova.virt.hardware [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 721.558076] env[63021]: DEBUG nova.virt.hardware [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 721.558258] env[63021]: DEBUG nova.virt.hardware [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 721.558937] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd0e0ae9-9b1f-4a3c-a92f-f996ae3672fa {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.567008] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de4a70d-f9a1-4892-bc39-fd4a978d5302 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.581700] env[63021]: ERROR nova.compute.manager [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f20d29e3-a559-4952-9e91-3afed97b91b6, please check neutron logs for more information. [ 721.581700] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] Traceback (most recent call last): [ 721.581700] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 721.581700] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] yield resources [ 721.581700] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 721.581700] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] self.driver.spawn(context, instance, image_meta, [ 721.581700] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 721.581700] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 721.581700] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 721.581700] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] vm_ref = self.build_virtual_machine(instance, [ 721.581700] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 721.582086] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] vif_infos = vmwarevif.get_vif_info(self._session, [ 721.582086] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 721.582086] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] for vif in network_info: [ 721.582086] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 721.582086] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] return self._sync_wrapper(fn, *args, **kwargs) [ 721.582086] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 721.582086] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] self.wait() [ 721.582086] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 721.582086] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] self[:] = self._gt.wait() [ 721.582086] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 721.582086] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] return self._exit_event.wait() [ 721.582086] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 721.582086] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] current.throw(*self._exc) [ 721.582512] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 721.582512] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] result = function(*args, **kwargs) [ 721.582512] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 721.582512] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] return func(*args, **kwargs) [ 721.582512] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 721.582512] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] raise e [ 721.582512] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 721.582512] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] nwinfo = self.network_api.allocate_for_instance( [ 721.582512] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 721.582512] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] created_port_ids = self._update_ports_for_instance( [ 721.582512] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 721.582512] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] with excutils.save_and_reraise_exception(): [ 721.582512] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.582947] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] self.force_reraise() [ 721.582947] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.582947] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] raise self.value [ 721.582947] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 721.582947] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] updated_port = self._update_port( [ 721.582947] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.582947] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] _ensure_no_port_binding_failure(port) [ 721.582947] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.582947] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] raise exception.PortBindingFailed(port_id=port['id']) [ 721.582947] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] nova.exception.PortBindingFailed: Binding failed for port f20d29e3-a559-4952-9e91-3afed97b91b6, please check neutron logs for more information. [ 721.582947] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] [ 721.582947] env[63021]: INFO nova.compute.manager [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Terminating instance [ 721.583982] env[63021]: DEBUG oslo_concurrency.lockutils [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Acquiring lock "refresh_cache-239d0087-462d-445a-a131-a23d4d15505f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.803102] env[63021]: DEBUG nova.network.neutron [req-6c2a07c3-0c35-49a9-aa68-d87f30f49053 req-82d8b557-6014-4f41-bc00-790af42986b4 service nova] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.873360] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.369s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.873894] env[63021]: DEBUG nova.compute.manager [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 721.876470] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.705s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.878550] env[63021]: INFO nova.compute.claims [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 721.882375] env[63021]: DEBUG oslo_concurrency.lockutils [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.909016] env[63021]: DEBUG nova.network.neutron [req-6c2a07c3-0c35-49a9-aa68-d87f30f49053 req-82d8b557-6014-4f41-bc00-790af42986b4 service nova] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.385630] env[63021]: DEBUG nova.compute.utils [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 722.387925] env[63021]: DEBUG nova.compute.manager [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 722.388030] env[63021]: DEBUG nova.network.neutron [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] [instance: c2523d70-12dd-444b-8839-b97378ef687e] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 722.414797] env[63021]: DEBUG oslo_concurrency.lockutils [req-6c2a07c3-0c35-49a9-aa68-d87f30f49053 req-82d8b557-6014-4f41-bc00-790af42986b4 service nova] Releasing lock "refresh_cache-239d0087-462d-445a-a131-a23d4d15505f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.415230] env[63021]: DEBUG oslo_concurrency.lockutils [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Acquired lock "refresh_cache-239d0087-462d-445a-a131-a23d4d15505f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.415308] env[63021]: DEBUG nova.network.neutron [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 722.455055] env[63021]: DEBUG nova.policy [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '255b89c53fa7496ea51c431c13964c75', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5c2ad1a14c7443d6ba737b736cca31eb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 722.861502] env[63021]: DEBUG nova.network.neutron [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Successfully created port: cfe90de3-370b-4f29-8bbc-e29a70448987 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 722.888421] env[63021]: DEBUG nova.compute.manager [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 722.940193] env[63021]: DEBUG nova.network.neutron [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 723.030166] env[63021]: DEBUG nova.network.neutron [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.247151] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3294592b-f2cb-4273-8b17-8ab8762f4e84 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.254899] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f074c63-1927-4b78-bcc7-1e312dfacb21 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.291425] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd0eadec-dc27-4c9f-bd97-3af22712acc8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.298526] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.298876] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.304925] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc0b8a3d-efa5-417e-9919-58d72b913e85 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.309470] env[63021]: DEBUG nova.compute.manager [req-ea7d62f7-592a-43e0-b5a7-ff904f7c5d96 req-6f7f8f2a-a023-4e10-9d5a-c394a0f0cf0a service nova] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Received event network-vif-deleted-f20d29e3-a559-4952-9e91-3afed97b91b6 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 723.320149] env[63021]: DEBUG nova.compute.provider_tree [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 723.538153] env[63021]: DEBUG oslo_concurrency.lockutils [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Releasing lock "refresh_cache-239d0087-462d-445a-a131-a23d4d15505f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.538153] env[63021]: DEBUG nova.compute.manager [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 723.538153] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 723.538153] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2e8e1f34-b614-4c71-a9f2-58edc8f1843c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.545884] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edf90eb4-e9b5-4745-a60c-abe700add654 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.565996] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 239d0087-462d-445a-a131-a23d4d15505f could not be found. [ 723.566218] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 723.566393] env[63021]: INFO nova.compute.manager [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Took 0.03 seconds to destroy the instance on the hypervisor. [ 723.566626] env[63021]: DEBUG oslo.service.loopingcall [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 723.566832] env[63021]: DEBUG nova.compute.manager [-] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 723.566927] env[63021]: DEBUG nova.network.neutron [-] [instance: 239d0087-462d-445a-a131-a23d4d15505f] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 723.582654] env[63021]: DEBUG nova.network.neutron [-] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 723.822894] env[63021]: DEBUG nova.scheduler.client.report [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 723.901810] env[63021]: DEBUG nova.compute.manager [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 723.929019] env[63021]: DEBUG nova.virt.hardware [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 723.929019] env[63021]: DEBUG nova.virt.hardware [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 723.929212] env[63021]: DEBUG nova.virt.hardware [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 723.929283] env[63021]: DEBUG nova.virt.hardware [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 723.929425] env[63021]: DEBUG nova.virt.hardware [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 723.929665] env[63021]: DEBUG nova.virt.hardware [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 723.929880] env[63021]: DEBUG nova.virt.hardware [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 723.930044] env[63021]: DEBUG nova.virt.hardware [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 723.930208] env[63021]: DEBUG nova.virt.hardware [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 723.930358] env[63021]: DEBUG nova.virt.hardware [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 723.930534] env[63021]: DEBUG nova.virt.hardware [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 723.931426] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-951081c6-d23f-4093-a27b-f15fbcda85bf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.939919] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d4949fa-0fad-42ed-97d0-7c46bc12a61b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.085288] env[63021]: DEBUG nova.network.neutron [-] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.158239] env[63021]: ERROR nova.compute.manager [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port cfe90de3-370b-4f29-8bbc-e29a70448987, please check neutron logs for more information. [ 724.158239] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 724.158239] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 724.158239] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 724.158239] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 724.158239] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 724.158239] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 724.158239] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 724.158239] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.158239] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 724.158239] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.158239] env[63021]: ERROR nova.compute.manager raise self.value [ 724.158239] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 724.158239] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 724.158239] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.158239] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 724.158813] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.158813] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 724.158813] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port cfe90de3-370b-4f29-8bbc-e29a70448987, please check neutron logs for more information. [ 724.158813] env[63021]: ERROR nova.compute.manager [ 724.158813] env[63021]: Traceback (most recent call last): [ 724.158813] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 724.158813] env[63021]: listener.cb(fileno) [ 724.158813] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 724.158813] env[63021]: result = function(*args, **kwargs) [ 724.158813] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 724.158813] env[63021]: return func(*args, **kwargs) [ 724.158813] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 724.158813] env[63021]: raise e [ 724.158813] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 724.158813] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 724.158813] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 724.158813] env[63021]: created_port_ids = self._update_ports_for_instance( [ 724.158813] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 724.158813] env[63021]: with excutils.save_and_reraise_exception(): [ 724.158813] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.158813] env[63021]: self.force_reraise() [ 724.158813] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.158813] env[63021]: raise self.value [ 724.158813] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 724.158813] env[63021]: updated_port = self._update_port( [ 724.158813] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.158813] env[63021]: _ensure_no_port_binding_failure(port) [ 724.158813] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.158813] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 724.159777] env[63021]: nova.exception.PortBindingFailed: Binding failed for port cfe90de3-370b-4f29-8bbc-e29a70448987, please check neutron logs for more information. [ 724.159777] env[63021]: Removing descriptor: 17 [ 724.159777] env[63021]: ERROR nova.compute.manager [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port cfe90de3-370b-4f29-8bbc-e29a70448987, please check neutron logs for more information. [ 724.159777] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] Traceback (most recent call last): [ 724.159777] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 724.159777] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] yield resources [ 724.159777] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 724.159777] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] self.driver.spawn(context, instance, image_meta, [ 724.159777] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 724.159777] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 724.159777] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 724.159777] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] vm_ref = self.build_virtual_machine(instance, [ 724.160188] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 724.160188] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] vif_infos = vmwarevif.get_vif_info(self._session, [ 724.160188] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 724.160188] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] for vif in network_info: [ 724.160188] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 724.160188] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] return self._sync_wrapper(fn, *args, **kwargs) [ 724.160188] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 724.160188] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] self.wait() [ 724.160188] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 724.160188] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] self[:] = self._gt.wait() [ 724.160188] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 724.160188] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] return self._exit_event.wait() [ 724.160188] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 724.160572] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] result = hub.switch() [ 724.160572] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 724.160572] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] return self.greenlet.switch() [ 724.160572] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 724.160572] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] result = function(*args, **kwargs) [ 724.160572] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 724.160572] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] return func(*args, **kwargs) [ 724.160572] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 724.160572] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] raise e [ 724.160572] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 724.160572] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] nwinfo = self.network_api.allocate_for_instance( [ 724.160572] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 724.160572] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] created_port_ids = self._update_ports_for_instance( [ 724.160952] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 724.160952] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] with excutils.save_and_reraise_exception(): [ 724.160952] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.160952] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] self.force_reraise() [ 724.160952] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.160952] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] raise self.value [ 724.160952] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 724.160952] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] updated_port = self._update_port( [ 724.160952] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.160952] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] _ensure_no_port_binding_failure(port) [ 724.160952] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.160952] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] raise exception.PortBindingFailed(port_id=port['id']) [ 724.161313] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] nova.exception.PortBindingFailed: Binding failed for port cfe90de3-370b-4f29-8bbc-e29a70448987, please check neutron logs for more information. [ 724.161313] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] [ 724.161313] env[63021]: INFO nova.compute.manager [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Terminating instance [ 724.161764] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Acquiring lock "refresh_cache-c2523d70-12dd-444b-8839-b97378ef687e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.161921] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Acquired lock "refresh_cache-c2523d70-12dd-444b-8839-b97378ef687e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.162096] env[63021]: DEBUG nova.network.neutron [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 724.328247] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.452s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.328800] env[63021]: DEBUG nova.compute.manager [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 724.331341] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.760s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.589049] env[63021]: INFO nova.compute.manager [-] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Took 1.02 seconds to deallocate network for instance. [ 724.591619] env[63021]: DEBUG nova.compute.claims [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 724.591787] env[63021]: DEBUG oslo_concurrency.lockutils [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.679432] env[63021]: DEBUG nova.network.neutron [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 724.749713] env[63021]: DEBUG nova.network.neutron [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.836397] env[63021]: DEBUG nova.compute.utils [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 724.841182] env[63021]: DEBUG nova.compute.manager [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 724.841586] env[63021]: DEBUG nova.network.neutron [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 724.881068] env[63021]: DEBUG nova.policy [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '445719c275e5469fb60006288d6e02f3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ae1fd9195eb04d068d317938f79c54f0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 725.162038] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aed63ef-e179-4a85-b36f-57fb453a196a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.170467] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67918371-5f9b-4165-bfa2-e09f2c04cd9f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.204278] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a90ae2e-6148-4bdb-ac5b-b54abb3ed01c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.211559] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeee8a32-591f-4868-9432-e5fe58fed080 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.227232] env[63021]: DEBUG nova.compute.provider_tree [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 725.252601] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Releasing lock "refresh_cache-c2523d70-12dd-444b-8839-b97378ef687e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.253208] env[63021]: DEBUG nova.compute.manager [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 725.253518] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 725.254208] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7d2a32a0-1245-4534-9de3-2edc1babe70c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.263913] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ef3de51-ed30-41cd-97a6-e34a5f079d2c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.286240] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c2523d70-12dd-444b-8839-b97378ef687e could not be found. [ 725.286459] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 725.286641] env[63021]: INFO nova.compute.manager [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Took 0.03 seconds to destroy the instance on the hypervisor. [ 725.286891] env[63021]: DEBUG oslo.service.loopingcall [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 725.287129] env[63021]: DEBUG nova.compute.manager [-] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 725.287223] env[63021]: DEBUG nova.network.neutron [-] [instance: c2523d70-12dd-444b-8839-b97378ef687e] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 725.296322] env[63021]: DEBUG nova.network.neutron [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Successfully created port: f0ca34f2-e0b1-42ec-8269-6bb0525ea8c0 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 725.305833] env[63021]: DEBUG nova.network.neutron [-] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.341955] env[63021]: DEBUG nova.compute.manager [req-e69bc043-fdf2-4af3-8452-6a3f6cfa5a6d req-6e2df57d-a29e-4213-bc44-ef1d92212b72 service nova] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Received event network-changed-cfe90de3-370b-4f29-8bbc-e29a70448987 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 725.342621] env[63021]: DEBUG nova.compute.manager [req-e69bc043-fdf2-4af3-8452-6a3f6cfa5a6d req-6e2df57d-a29e-4213-bc44-ef1d92212b72 service nova] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Refreshing instance network info cache due to event network-changed-cfe90de3-370b-4f29-8bbc-e29a70448987. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 725.342786] env[63021]: DEBUG oslo_concurrency.lockutils [req-e69bc043-fdf2-4af3-8452-6a3f6cfa5a6d req-6e2df57d-a29e-4213-bc44-ef1d92212b72 service nova] Acquiring lock "refresh_cache-c2523d70-12dd-444b-8839-b97378ef687e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.342879] env[63021]: DEBUG oslo_concurrency.lockutils [req-e69bc043-fdf2-4af3-8452-6a3f6cfa5a6d req-6e2df57d-a29e-4213-bc44-ef1d92212b72 service nova] Acquired lock "refresh_cache-c2523d70-12dd-444b-8839-b97378ef687e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.343448] env[63021]: DEBUG nova.network.neutron [req-e69bc043-fdf2-4af3-8452-6a3f6cfa5a6d req-6e2df57d-a29e-4213-bc44-ef1d92212b72 service nova] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Refreshing network info cache for port cfe90de3-370b-4f29-8bbc-e29a70448987 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 725.345363] env[63021]: DEBUG nova.compute.manager [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 725.731189] env[63021]: DEBUG nova.scheduler.client.report [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 725.811877] env[63021]: DEBUG nova.network.neutron [-] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.869985] env[63021]: DEBUG nova.network.neutron [req-e69bc043-fdf2-4af3-8452-6a3f6cfa5a6d req-6e2df57d-a29e-4213-bc44-ef1d92212b72 service nova] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.963278] env[63021]: DEBUG nova.network.neutron [req-e69bc043-fdf2-4af3-8452-6a3f6cfa5a6d req-6e2df57d-a29e-4213-bc44-ef1d92212b72 service nova] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.241073] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.910s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.241697] env[63021]: ERROR nova.compute.manager [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a1802ba1-ac36-4a83-a474-7c6e6253ca38, please check neutron logs for more information. [ 726.241697] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Traceback (most recent call last): [ 726.241697] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 726.241697] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] self.driver.spawn(context, instance, image_meta, [ 726.241697] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 726.241697] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] self._vmops.spawn(context, instance, image_meta, injected_files, [ 726.241697] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 726.241697] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] vm_ref = self.build_virtual_machine(instance, [ 726.241697] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 726.241697] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] vif_infos = vmwarevif.get_vif_info(self._session, [ 726.241697] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 726.241985] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] for vif in network_info: [ 726.241985] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 726.241985] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] return self._sync_wrapper(fn, *args, **kwargs) [ 726.241985] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 726.241985] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] self.wait() [ 726.241985] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 726.241985] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] self[:] = self._gt.wait() [ 726.241985] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 726.241985] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] return self._exit_event.wait() [ 726.241985] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 726.241985] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] current.throw(*self._exc) [ 726.241985] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.241985] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] result = function(*args, **kwargs) [ 726.242355] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 726.242355] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] return func(*args, **kwargs) [ 726.242355] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 726.242355] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] raise e [ 726.242355] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.242355] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] nwinfo = self.network_api.allocate_for_instance( [ 726.242355] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 726.242355] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] created_port_ids = self._update_ports_for_instance( [ 726.242355] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 726.242355] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] with excutils.save_and_reraise_exception(): [ 726.242355] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.242355] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] self.force_reraise() [ 726.242355] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.242715] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] raise self.value [ 726.242715] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 726.242715] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] updated_port = self._update_port( [ 726.242715] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.242715] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] _ensure_no_port_binding_failure(port) [ 726.242715] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.242715] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] raise exception.PortBindingFailed(port_id=port['id']) [ 726.242715] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] nova.exception.PortBindingFailed: Binding failed for port a1802ba1-ac36-4a83-a474-7c6e6253ca38, please check neutron logs for more information. [ 726.242715] env[63021]: ERROR nova.compute.manager [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] [ 726.242715] env[63021]: DEBUG nova.compute.utils [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Binding failed for port a1802ba1-ac36-4a83-a474-7c6e6253ca38, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 726.243819] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.300s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.246817] env[63021]: DEBUG nova.compute.manager [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Build of instance aef64522-2dbb-4bc5-84b2-a0bf8a108574 was re-scheduled: Binding failed for port a1802ba1-ac36-4a83-a474-7c6e6253ca38, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 726.247276] env[63021]: DEBUG nova.compute.manager [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 726.247446] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Acquiring lock "refresh_cache-aef64522-2dbb-4bc5-84b2-a0bf8a108574" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.247589] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Acquired lock "refresh_cache-aef64522-2dbb-4bc5-84b2-a0bf8a108574" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.247743] env[63021]: DEBUG nova.network.neutron [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 726.314342] env[63021]: INFO nova.compute.manager [-] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Took 1.03 seconds to deallocate network for instance. [ 726.316550] env[63021]: DEBUG nova.compute.claims [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 726.316723] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.354950] env[63021]: DEBUG nova.compute.manager [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 726.368491] env[63021]: ERROR nova.compute.manager [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f0ca34f2-e0b1-42ec-8269-6bb0525ea8c0, please check neutron logs for more information. [ 726.368491] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 726.368491] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.368491] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 726.368491] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 726.368491] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 726.368491] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 726.368491] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 726.368491] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.368491] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 726.368491] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.368491] env[63021]: ERROR nova.compute.manager raise self.value [ 726.368491] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 726.368491] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 726.368491] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.368491] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 726.369114] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.369114] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 726.369114] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f0ca34f2-e0b1-42ec-8269-6bb0525ea8c0, please check neutron logs for more information. [ 726.369114] env[63021]: ERROR nova.compute.manager [ 726.369114] env[63021]: Traceback (most recent call last): [ 726.369114] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 726.369114] env[63021]: listener.cb(fileno) [ 726.369114] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.369114] env[63021]: result = function(*args, **kwargs) [ 726.369114] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 726.369114] env[63021]: return func(*args, **kwargs) [ 726.369114] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 726.369114] env[63021]: raise e [ 726.369114] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.369114] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 726.369114] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 726.369114] env[63021]: created_port_ids = self._update_ports_for_instance( [ 726.369114] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 726.369114] env[63021]: with excutils.save_and_reraise_exception(): [ 726.369114] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.369114] env[63021]: self.force_reraise() [ 726.369114] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.369114] env[63021]: raise self.value [ 726.369114] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 726.369114] env[63021]: updated_port = self._update_port( [ 726.369114] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.369114] env[63021]: _ensure_no_port_binding_failure(port) [ 726.369114] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.369114] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 726.371209] env[63021]: nova.exception.PortBindingFailed: Binding failed for port f0ca34f2-e0b1-42ec-8269-6bb0525ea8c0, please check neutron logs for more information. [ 726.371209] env[63021]: Removing descriptor: 17 [ 726.384600] env[63021]: DEBUG nova.virt.hardware [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 726.384844] env[63021]: DEBUG nova.virt.hardware [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 726.384989] env[63021]: DEBUG nova.virt.hardware [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 726.385182] env[63021]: DEBUG nova.virt.hardware [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 726.385325] env[63021]: DEBUG nova.virt.hardware [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 726.385470] env[63021]: DEBUG nova.virt.hardware [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 726.385670] env[63021]: DEBUG nova.virt.hardware [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 726.385823] env[63021]: DEBUG nova.virt.hardware [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 726.385984] env[63021]: DEBUG nova.virt.hardware [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 726.386160] env[63021]: DEBUG nova.virt.hardware [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 726.386327] env[63021]: DEBUG nova.virt.hardware [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 726.387180] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-225b64d9-4e01-49d0-b0de-8257c98ca57e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.396643] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5f5cfc4-11ab-4ed7-b3ae-d33746816844 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.410008] env[63021]: ERROR nova.compute.manager [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f0ca34f2-e0b1-42ec-8269-6bb0525ea8c0, please check neutron logs for more information. [ 726.410008] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Traceback (most recent call last): [ 726.410008] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 726.410008] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] yield resources [ 726.410008] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 726.410008] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] self.driver.spawn(context, instance, image_meta, [ 726.410008] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 726.410008] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] self._vmops.spawn(context, instance, image_meta, injected_files, [ 726.410008] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 726.410008] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] vm_ref = self.build_virtual_machine(instance, [ 726.410008] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 726.410462] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] vif_infos = vmwarevif.get_vif_info(self._session, [ 726.410462] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 726.410462] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] for vif in network_info: [ 726.410462] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 726.410462] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] return self._sync_wrapper(fn, *args, **kwargs) [ 726.410462] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 726.410462] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] self.wait() [ 726.410462] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 726.410462] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] self[:] = self._gt.wait() [ 726.410462] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 726.410462] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] return self._exit_event.wait() [ 726.410462] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 726.410462] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] current.throw(*self._exc) [ 726.410886] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.410886] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] result = function(*args, **kwargs) [ 726.410886] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 726.410886] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] return func(*args, **kwargs) [ 726.410886] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 726.410886] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] raise e [ 726.410886] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.410886] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] nwinfo = self.network_api.allocate_for_instance( [ 726.410886] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 726.410886] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] created_port_ids = self._update_ports_for_instance( [ 726.410886] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 726.410886] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] with excutils.save_and_reraise_exception(): [ 726.410886] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.411295] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] self.force_reraise() [ 726.411295] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.411295] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] raise self.value [ 726.411295] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 726.411295] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] updated_port = self._update_port( [ 726.411295] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.411295] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] _ensure_no_port_binding_failure(port) [ 726.411295] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.411295] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] raise exception.PortBindingFailed(port_id=port['id']) [ 726.411295] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] nova.exception.PortBindingFailed: Binding failed for port f0ca34f2-e0b1-42ec-8269-6bb0525ea8c0, please check neutron logs for more information. [ 726.411295] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] [ 726.411295] env[63021]: INFO nova.compute.manager [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Terminating instance [ 726.412259] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Acquiring lock "refresh_cache-26896531-07eb-4e44-8dd8-e5237bfd7c39" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.412443] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Acquired lock "refresh_cache-26896531-07eb-4e44-8dd8-e5237bfd7c39" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.412609] env[63021]: DEBUG nova.network.neutron [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 726.466039] env[63021]: DEBUG oslo_concurrency.lockutils [req-e69bc043-fdf2-4af3-8452-6a3f6cfa5a6d req-6e2df57d-a29e-4213-bc44-ef1d92212b72 service nova] Releasing lock "refresh_cache-c2523d70-12dd-444b-8839-b97378ef687e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.466218] env[63021]: DEBUG nova.compute.manager [req-e69bc043-fdf2-4af3-8452-6a3f6cfa5a6d req-6e2df57d-a29e-4213-bc44-ef1d92212b72 service nova] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Received event network-vif-deleted-cfe90de3-370b-4f29-8bbc-e29a70448987 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 726.766706] env[63021]: DEBUG nova.network.neutron [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.841087] env[63021]: DEBUG nova.network.neutron [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.938656] env[63021]: DEBUG nova.network.neutron [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.062248] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e40051ca-768c-4b0f-b641-1f60a3020e08 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.069703] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c48842e7-17c2-4028-94f4-896e31ca08ea {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.073725] env[63021]: DEBUG nova.network.neutron [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.101334] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81b3ecdd-5d4a-4238-a279-475c4bc136dd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.109072] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3158fb2-58a8-4a41-872c-1ab4411384d5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.122484] env[63021]: DEBUG nova.compute.provider_tree [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 727.343961] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Releasing lock "refresh_cache-aef64522-2dbb-4bc5-84b2-a0bf8a108574" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.344187] env[63021]: DEBUG nova.compute.manager [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 727.344374] env[63021]: DEBUG nova.compute.manager [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 727.344545] env[63021]: DEBUG nova.network.neutron [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 727.370017] env[63021]: DEBUG nova.network.neutron [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.376182] env[63021]: DEBUG nova.compute.manager [req-4d1fb229-22ed-4871-a436-a771723a6cc1 req-6d56ff25-ccfa-4f22-91af-8b04332cc091 service nova] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Received event network-changed-f0ca34f2-e0b1-42ec-8269-6bb0525ea8c0 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 727.376371] env[63021]: DEBUG nova.compute.manager [req-4d1fb229-22ed-4871-a436-a771723a6cc1 req-6d56ff25-ccfa-4f22-91af-8b04332cc091 service nova] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Refreshing instance network info cache due to event network-changed-f0ca34f2-e0b1-42ec-8269-6bb0525ea8c0. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 727.376545] env[63021]: DEBUG oslo_concurrency.lockutils [req-4d1fb229-22ed-4871-a436-a771723a6cc1 req-6d56ff25-ccfa-4f22-91af-8b04332cc091 service nova] Acquiring lock "refresh_cache-26896531-07eb-4e44-8dd8-e5237bfd7c39" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.578343] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Releasing lock "refresh_cache-26896531-07eb-4e44-8dd8-e5237bfd7c39" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.578768] env[63021]: DEBUG nova.compute.manager [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 727.578959] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 727.579283] env[63021]: DEBUG oslo_concurrency.lockutils [req-4d1fb229-22ed-4871-a436-a771723a6cc1 req-6d56ff25-ccfa-4f22-91af-8b04332cc091 service nova] Acquired lock "refresh_cache-26896531-07eb-4e44-8dd8-e5237bfd7c39" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.579449] env[63021]: DEBUG nova.network.neutron [req-4d1fb229-22ed-4871-a436-a771723a6cc1 req-6d56ff25-ccfa-4f22-91af-8b04332cc091 service nova] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Refreshing network info cache for port f0ca34f2-e0b1-42ec-8269-6bb0525ea8c0 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 727.580544] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0567420c-261e-4407-9bac-2855c2089682 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.590442] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1830e99-c23a-4a74-ad35-633ecc1b4559 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.611998] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 26896531-07eb-4e44-8dd8-e5237bfd7c39 could not be found. [ 727.612236] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 727.612458] env[63021]: INFO nova.compute.manager [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Took 0.03 seconds to destroy the instance on the hypervisor. [ 727.612752] env[63021]: DEBUG oslo.service.loopingcall [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 727.612921] env[63021]: DEBUG nova.compute.manager [-] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 727.613018] env[63021]: DEBUG nova.network.neutron [-] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 727.624858] env[63021]: DEBUG nova.scheduler.client.report [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 727.634297] env[63021]: DEBUG nova.network.neutron [-] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.875588] env[63021]: DEBUG nova.network.neutron [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.098700] env[63021]: DEBUG nova.network.neutron [req-4d1fb229-22ed-4871-a436-a771723a6cc1 req-6d56ff25-ccfa-4f22-91af-8b04332cc091 service nova] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.132470] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.889s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.133177] env[63021]: ERROR nova.compute.manager [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ea5445cf-f96e-48c9-8ee4-e1e62207c32e, please check neutron logs for more information. [ 728.133177] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Traceback (most recent call last): [ 728.133177] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 728.133177] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] self.driver.spawn(context, instance, image_meta, [ 728.133177] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 728.133177] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] self._vmops.spawn(context, instance, image_meta, injected_files, [ 728.133177] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 728.133177] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] vm_ref = self.build_virtual_machine(instance, [ 728.133177] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 728.133177] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] vif_infos = vmwarevif.get_vif_info(self._session, [ 728.133177] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 728.133554] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] for vif in network_info: [ 728.133554] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 728.133554] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] return self._sync_wrapper(fn, *args, **kwargs) [ 728.133554] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 728.133554] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] self.wait() [ 728.133554] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 728.133554] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] self[:] = self._gt.wait() [ 728.133554] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 728.133554] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] return self._exit_event.wait() [ 728.133554] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 728.133554] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] result = hub.switch() [ 728.133554] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 728.133554] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] return self.greenlet.switch() [ 728.133966] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 728.133966] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] result = function(*args, **kwargs) [ 728.133966] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 728.133966] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] return func(*args, **kwargs) [ 728.133966] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 728.133966] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] raise e [ 728.133966] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 728.133966] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] nwinfo = self.network_api.allocate_for_instance( [ 728.133966] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 728.133966] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] created_port_ids = self._update_ports_for_instance( [ 728.133966] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 728.133966] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] with excutils.save_and_reraise_exception(): [ 728.133966] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.134366] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] self.force_reraise() [ 728.134366] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.134366] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] raise self.value [ 728.134366] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 728.134366] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] updated_port = self._update_port( [ 728.134366] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.134366] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] _ensure_no_port_binding_failure(port) [ 728.134366] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.134366] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] raise exception.PortBindingFailed(port_id=port['id']) [ 728.134366] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] nova.exception.PortBindingFailed: Binding failed for port ea5445cf-f96e-48c9-8ee4-e1e62207c32e, please check neutron logs for more information. [ 728.134366] env[63021]: ERROR nova.compute.manager [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] [ 728.134725] env[63021]: DEBUG nova.compute.utils [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Binding failed for port ea5445cf-f96e-48c9-8ee4-e1e62207c32e, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 728.135317] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.097s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.137954] env[63021]: DEBUG nova.compute.manager [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Build of instance c6ed1f3b-e80e-4545-8338-8e3aead5ba16 was re-scheduled: Binding failed for port ea5445cf-f96e-48c9-8ee4-e1e62207c32e, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 728.138320] env[63021]: DEBUG nova.compute.manager [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 728.138543] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Acquiring lock "refresh_cache-c6ed1f3b-e80e-4545-8338-8e3aead5ba16" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.138686] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Acquired lock "refresh_cache-c6ed1f3b-e80e-4545-8338-8e3aead5ba16" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.138841] env[63021]: DEBUG nova.network.neutron [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 728.139670] env[63021]: DEBUG nova.network.neutron [-] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.204257] env[63021]: DEBUG nova.network.neutron [req-4d1fb229-22ed-4871-a436-a771723a6cc1 req-6d56ff25-ccfa-4f22-91af-8b04332cc091 service nova] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.379010] env[63021]: INFO nova.compute.manager [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] [instance: aef64522-2dbb-4bc5-84b2-a0bf8a108574] Took 1.03 seconds to deallocate network for instance. [ 728.646381] env[63021]: INFO nova.compute.manager [-] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Took 1.03 seconds to deallocate network for instance. [ 728.649279] env[63021]: DEBUG nova.compute.claims [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 728.649279] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.659468] env[63021]: DEBUG nova.network.neutron [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.706118] env[63021]: DEBUG oslo_concurrency.lockutils [req-4d1fb229-22ed-4871-a436-a771723a6cc1 req-6d56ff25-ccfa-4f22-91af-8b04332cc091 service nova] Releasing lock "refresh_cache-26896531-07eb-4e44-8dd8-e5237bfd7c39" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.706365] env[63021]: DEBUG nova.compute.manager [req-4d1fb229-22ed-4871-a436-a771723a6cc1 req-6d56ff25-ccfa-4f22-91af-8b04332cc091 service nova] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Received event network-vif-deleted-f0ca34f2-e0b1-42ec-8269-6bb0525ea8c0 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 728.750818] env[63021]: DEBUG nova.network.neutron [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.918616] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f652f37b-558b-4588-b226-6bb779c7a917 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.927388] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cf8202f-8347-4114-96ca-f68eba9b855f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.959046] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0fc29f4-56cc-4615-84c0-0e7876eadce7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.966407] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b4069cd-84b7-4e8c-9ed8-7df3a3469a13 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.979382] env[63021]: DEBUG nova.compute.provider_tree [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 729.254822] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Releasing lock "refresh_cache-c6ed1f3b-e80e-4545-8338-8e3aead5ba16" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.255154] env[63021]: DEBUG nova.compute.manager [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 729.255472] env[63021]: DEBUG nova.compute.manager [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 729.255699] env[63021]: DEBUG nova.network.neutron [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 729.275667] env[63021]: DEBUG nova.network.neutron [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.406730] env[63021]: INFO nova.scheduler.client.report [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Deleted allocations for instance aef64522-2dbb-4bc5-84b2-a0bf8a108574 [ 729.482217] env[63021]: DEBUG nova.scheduler.client.report [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 729.778601] env[63021]: DEBUG nova.network.neutron [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.916840] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6f537925-1564-4d0a-97d4-f3b503670f65 tempest-ServersTestBootFromVolume-1057289707 tempest-ServersTestBootFromVolume-1057289707-project-member] Lock "aef64522-2dbb-4bc5-84b2-a0bf8a108574" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 135.614s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.987640] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.852s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.988559] env[63021]: ERROR nova.compute.manager [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5b7f5119-dade-4047-86cc-6cffab3760ad, please check neutron logs for more information. [ 729.988559] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Traceback (most recent call last): [ 729.988559] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 729.988559] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] self.driver.spawn(context, instance, image_meta, [ 729.988559] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 729.988559] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 729.988559] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 729.988559] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] vm_ref = self.build_virtual_machine(instance, [ 729.988559] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 729.988559] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] vif_infos = vmwarevif.get_vif_info(self._session, [ 729.988559] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 729.988998] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] for vif in network_info: [ 729.988998] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 729.988998] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] return self._sync_wrapper(fn, *args, **kwargs) [ 729.988998] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 729.988998] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] self.wait() [ 729.988998] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 729.988998] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] self[:] = self._gt.wait() [ 729.988998] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 729.988998] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] return self._exit_event.wait() [ 729.988998] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 729.988998] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] result = hub.switch() [ 729.988998] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 729.988998] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] return self.greenlet.switch() [ 729.989404] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 729.989404] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] result = function(*args, **kwargs) [ 729.989404] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 729.989404] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] return func(*args, **kwargs) [ 729.989404] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 729.989404] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] raise e [ 729.989404] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 729.989404] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] nwinfo = self.network_api.allocate_for_instance( [ 729.989404] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 729.989404] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] created_port_ids = self._update_ports_for_instance( [ 729.989404] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 729.989404] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] with excutils.save_and_reraise_exception(): [ 729.989404] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 729.989785] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] self.force_reraise() [ 729.989785] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 729.989785] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] raise self.value [ 729.989785] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 729.989785] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] updated_port = self._update_port( [ 729.989785] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 729.989785] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] _ensure_no_port_binding_failure(port) [ 729.989785] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 729.989785] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] raise exception.PortBindingFailed(port_id=port['id']) [ 729.989785] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] nova.exception.PortBindingFailed: Binding failed for port 5b7f5119-dade-4047-86cc-6cffab3760ad, please check neutron logs for more information. [ 729.989785] env[63021]: ERROR nova.compute.manager [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] [ 729.990170] env[63021]: DEBUG nova.compute.utils [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Binding failed for port 5b7f5119-dade-4047-86cc-6cffab3760ad, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 729.990209] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.227s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.992970] env[63021]: DEBUG nova.compute.manager [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Build of instance d36a9ed2-6849-43fc-ab88-77468bde337f was re-scheduled: Binding failed for port 5b7f5119-dade-4047-86cc-6cffab3760ad, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 729.993389] env[63021]: DEBUG nova.compute.manager [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 729.993655] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Acquiring lock "refresh_cache-d36a9ed2-6849-43fc-ab88-77468bde337f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.993807] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Acquired lock "refresh_cache-d36a9ed2-6849-43fc-ab88-77468bde337f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.993966] env[63021]: DEBUG nova.network.neutron [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 730.282030] env[63021]: INFO nova.compute.manager [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: c6ed1f3b-e80e-4545-8338-8e3aead5ba16] Took 1.03 seconds to deallocate network for instance. [ 730.420074] env[63021]: DEBUG nova.compute.manager [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 730.516928] env[63021]: DEBUG nova.network.neutron [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 730.642891] env[63021]: DEBUG nova.network.neutron [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.795121] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e04ab528-9ade-46df-8234-ff90d28e7106 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.808627] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ea77982-1bd8-4882-aa02-836eac530f1c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.846286] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c48569d-7c9f-4974-ba6a-83bb5997e6e0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.854025] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eab2f34-a6ee-4b79-9720-71e23ba38a39 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.871058] env[63021]: DEBUG nova.compute.provider_tree [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 730.939420] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.146983] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Releasing lock "refresh_cache-d36a9ed2-6849-43fc-ab88-77468bde337f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.146983] env[63021]: DEBUG nova.compute.manager [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 731.146983] env[63021]: DEBUG nova.compute.manager [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 731.146983] env[63021]: DEBUG nova.network.neutron [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 731.162996] env[63021]: DEBUG nova.network.neutron [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.321384] env[63021]: INFO nova.scheduler.client.report [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Deleted allocations for instance c6ed1f3b-e80e-4545-8338-8e3aead5ba16 [ 731.375687] env[63021]: DEBUG nova.scheduler.client.report [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 731.666093] env[63021]: DEBUG nova.network.neutron [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.830044] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dd9008e2-8661-4dd5-9eeb-cd7ffd67cf40 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Lock "c6ed1f3b-e80e-4545-8338-8e3aead5ba16" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.347s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.879186] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.889s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.879831] env[63021]: ERROR nova.compute.manager [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 53c3f4ae-ad6b-431d-8d8c-50e340518f90, please check neutron logs for more information. [ 731.879831] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Traceback (most recent call last): [ 731.879831] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 731.879831] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] self.driver.spawn(context, instance, image_meta, [ 731.879831] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 731.879831] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 731.879831] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 731.879831] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] vm_ref = self.build_virtual_machine(instance, [ 731.879831] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 731.879831] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] vif_infos = vmwarevif.get_vif_info(self._session, [ 731.879831] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 731.880395] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] for vif in network_info: [ 731.880395] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 731.880395] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] return self._sync_wrapper(fn, *args, **kwargs) [ 731.880395] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 731.880395] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] self.wait() [ 731.880395] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 731.880395] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] self[:] = self._gt.wait() [ 731.880395] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 731.880395] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] return self._exit_event.wait() [ 731.880395] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 731.880395] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] result = hub.switch() [ 731.880395] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 731.880395] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] return self.greenlet.switch() [ 731.881028] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 731.881028] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] result = function(*args, **kwargs) [ 731.881028] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 731.881028] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] return func(*args, **kwargs) [ 731.881028] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 731.881028] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] raise e [ 731.881028] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 731.881028] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] nwinfo = self.network_api.allocate_for_instance( [ 731.881028] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 731.881028] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] created_port_ids = self._update_ports_for_instance( [ 731.881028] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 731.881028] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] with excutils.save_and_reraise_exception(): [ 731.881028] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 731.881590] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] self.force_reraise() [ 731.881590] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 731.881590] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] raise self.value [ 731.881590] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 731.881590] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] updated_port = self._update_port( [ 731.881590] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 731.881590] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] _ensure_no_port_binding_failure(port) [ 731.881590] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 731.881590] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] raise exception.PortBindingFailed(port_id=port['id']) [ 731.881590] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] nova.exception.PortBindingFailed: Binding failed for port 53c3f4ae-ad6b-431d-8d8c-50e340518f90, please check neutron logs for more information. [ 731.881590] env[63021]: ERROR nova.compute.manager [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] [ 731.881925] env[63021]: DEBUG nova.compute.utils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Binding failed for port 53c3f4ae-ad6b-431d-8d8c-50e340518f90, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 731.881925] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.857s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.885950] env[63021]: DEBUG nova.compute.manager [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Build of instance 476e1e73-7050-4e53-bff8-d5d9f0982f1a was re-scheduled: Binding failed for port 53c3f4ae-ad6b-431d-8d8c-50e340518f90, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 731.886491] env[63021]: DEBUG nova.compute.manager [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 731.886662] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquiring lock "refresh_cache-476e1e73-7050-4e53-bff8-d5d9f0982f1a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.886804] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquired lock "refresh_cache-476e1e73-7050-4e53-bff8-d5d9f0982f1a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.886957] env[63021]: DEBUG nova.network.neutron [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 732.169258] env[63021]: INFO nova.compute.manager [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] [instance: d36a9ed2-6849-43fc-ab88-77468bde337f] Took 1.02 seconds to deallocate network for instance. [ 732.332704] env[63021]: DEBUG nova.compute.manager [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 732.436058] env[63021]: DEBUG nova.network.neutron [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.717627] env[63021]: DEBUG nova.network.neutron [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.741528] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fe2580c-a40f-4b75-b3f0-25c867e627f3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.749931] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c787bd5e-8e00-45aa-9e75-ba1a891db0fe {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.783536] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91484e7d-8111-44c7-9c82-d896059f6a7e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.791801] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53c5e5cd-0b40-40b6-8812-c14443ff8c5a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.806837] env[63021]: DEBUG nova.compute.provider_tree [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 732.851294] env[63021]: DEBUG oslo_concurrency.lockutils [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.144870] env[63021]: DEBUG oslo_concurrency.lockutils [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Acquiring lock "868b1aeb-8109-4cb9-bb58-33f3d92989ad" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.145103] env[63021]: DEBUG oslo_concurrency.lockutils [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Lock "868b1aeb-8109-4cb9-bb58-33f3d92989ad" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.202270] env[63021]: INFO nova.scheduler.client.report [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Deleted allocations for instance d36a9ed2-6849-43fc-ab88-77468bde337f [ 733.221193] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Releasing lock "refresh_cache-476e1e73-7050-4e53-bff8-d5d9f0982f1a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.221535] env[63021]: DEBUG nova.compute.manager [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 733.221636] env[63021]: DEBUG nova.compute.manager [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 733.221757] env[63021]: DEBUG nova.network.neutron [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 733.238338] env[63021]: DEBUG nova.network.neutron [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.310227] env[63021]: DEBUG nova.scheduler.client.report [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 733.710915] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b43ad035-ef68-42e6-af1a-ab82907f1346 tempest-ServerAddressesNegativeTestJSON-835801917 tempest-ServerAddressesNegativeTestJSON-835801917-project-member] Lock "d36a9ed2-6849-43fc-ab88-77468bde337f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.187s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.740941] env[63021]: DEBUG nova.network.neutron [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.815401] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.933s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.816092] env[63021]: ERROR nova.compute.manager [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 624e089c-f559-40e6-b38b-8abf34ede8a9, please check neutron logs for more information. [ 733.816092] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Traceback (most recent call last): [ 733.816092] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 733.816092] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] self.driver.spawn(context, instance, image_meta, [ 733.816092] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 733.816092] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 733.816092] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 733.816092] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] vm_ref = self.build_virtual_machine(instance, [ 733.816092] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 733.816092] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] vif_infos = vmwarevif.get_vif_info(self._session, [ 733.816092] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 733.816455] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] for vif in network_info: [ 733.816455] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 733.816455] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] return self._sync_wrapper(fn, *args, **kwargs) [ 733.816455] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 733.816455] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] self.wait() [ 733.816455] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 733.816455] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] self[:] = self._gt.wait() [ 733.816455] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 733.816455] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] return self._exit_event.wait() [ 733.816455] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 733.816455] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] current.throw(*self._exc) [ 733.816455] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 733.816455] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] result = function(*args, **kwargs) [ 733.816865] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 733.816865] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] return func(*args, **kwargs) [ 733.816865] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 733.816865] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] raise e [ 733.816865] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 733.816865] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] nwinfo = self.network_api.allocate_for_instance( [ 733.816865] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 733.816865] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] created_port_ids = self._update_ports_for_instance( [ 733.816865] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 733.816865] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] with excutils.save_and_reraise_exception(): [ 733.816865] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.816865] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] self.force_reraise() [ 733.816865] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.817267] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] raise self.value [ 733.817267] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 733.817267] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] updated_port = self._update_port( [ 733.817267] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 733.817267] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] _ensure_no_port_binding_failure(port) [ 733.817267] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 733.817267] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] raise exception.PortBindingFailed(port_id=port['id']) [ 733.817267] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] nova.exception.PortBindingFailed: Binding failed for port 624e089c-f559-40e6-b38b-8abf34ede8a9, please check neutron logs for more information. [ 733.817267] env[63021]: ERROR nova.compute.manager [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] [ 733.817267] env[63021]: DEBUG nova.compute.utils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Binding failed for port 624e089c-f559-40e6-b38b-8abf34ede8a9, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 733.817999] env[63021]: DEBUG oslo_concurrency.lockutils [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.236s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.821405] env[63021]: DEBUG nova.compute.manager [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Build of instance 28d70e1d-da2a-4c9c-b297-86d026ce42c3 was re-scheduled: Binding failed for port 624e089c-f559-40e6-b38b-8abf34ede8a9, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 733.821851] env[63021]: DEBUG nova.compute.manager [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 733.822087] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquiring lock "refresh_cache-28d70e1d-da2a-4c9c-b297-86d026ce42c3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.822234] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquired lock "refresh_cache-28d70e1d-da2a-4c9c-b297-86d026ce42c3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.822568] env[63021]: DEBUG nova.network.neutron [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 734.214127] env[63021]: DEBUG nova.compute.manager [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 734.243835] env[63021]: INFO nova.compute.manager [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 476e1e73-7050-4e53-bff8-d5d9f0982f1a] Took 1.02 seconds to deallocate network for instance. [ 734.346944] env[63021]: DEBUG nova.network.neutron [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 734.496668] env[63021]: DEBUG nova.network.neutron [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.682027] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06b56de4-5205-4d7b-8de0-a423a43ea257 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.690661] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c71948eb-6c28-4274-bad4-c500139ca2b1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.724982] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cef0b010-1dcc-4380-a206-31b6d6049e3d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.732780] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05b8f263-ff0d-4521-9e2b-4948bbf4b707 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.745904] env[63021]: DEBUG nova.compute.provider_tree [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 734.747706] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.974125] env[63021]: DEBUG oslo_concurrency.lockutils [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquiring lock "df0b6974-7f78-4b73-8583-d09754fbacea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.974125] env[63021]: DEBUG oslo_concurrency.lockutils [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Lock "df0b6974-7f78-4b73-8583-d09754fbacea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.000157] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Releasing lock "refresh_cache-28d70e1d-da2a-4c9c-b297-86d026ce42c3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.000157] env[63021]: DEBUG nova.compute.manager [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 735.000157] env[63021]: DEBUG nova.compute.manager [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 735.000157] env[63021]: DEBUG nova.network.neutron [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 735.032036] env[63021]: DEBUG nova.network.neutron [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.252693] env[63021]: DEBUG nova.scheduler.client.report [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 735.285335] env[63021]: INFO nova.scheduler.client.report [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Deleted allocations for instance 476e1e73-7050-4e53-bff8-d5d9f0982f1a [ 735.537235] env[63021]: DEBUG nova.network.neutron [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.759927] env[63021]: DEBUG oslo_concurrency.lockutils [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.942s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.760632] env[63021]: ERROR nova.compute.manager [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9c006af2-d842-42f6-9191-74570d1de950, please check neutron logs for more information. [ 735.760632] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Traceback (most recent call last): [ 735.760632] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 735.760632] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] self.driver.spawn(context, instance, image_meta, [ 735.760632] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 735.760632] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 735.760632] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 735.760632] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] vm_ref = self.build_virtual_machine(instance, [ 735.760632] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 735.760632] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] vif_infos = vmwarevif.get_vif_info(self._session, [ 735.760632] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 735.760966] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] for vif in network_info: [ 735.760966] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 735.760966] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] return self._sync_wrapper(fn, *args, **kwargs) [ 735.760966] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 735.760966] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] self.wait() [ 735.760966] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 735.760966] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] self[:] = self._gt.wait() [ 735.760966] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 735.760966] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] return self._exit_event.wait() [ 735.760966] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 735.760966] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] result = hub.switch() [ 735.760966] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 735.760966] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] return self.greenlet.switch() [ 735.761313] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 735.761313] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] result = function(*args, **kwargs) [ 735.761313] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 735.761313] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] return func(*args, **kwargs) [ 735.761313] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 735.761313] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] raise e [ 735.761313] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 735.761313] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] nwinfo = self.network_api.allocate_for_instance( [ 735.761313] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 735.761313] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] created_port_ids = self._update_ports_for_instance( [ 735.761313] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 735.761313] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] with excutils.save_and_reraise_exception(): [ 735.761313] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.761713] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] self.force_reraise() [ 735.761713] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.761713] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] raise self.value [ 735.761713] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 735.761713] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] updated_port = self._update_port( [ 735.761713] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.761713] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] _ensure_no_port_binding_failure(port) [ 735.761713] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.761713] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] raise exception.PortBindingFailed(port_id=port['id']) [ 735.761713] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] nova.exception.PortBindingFailed: Binding failed for port 9c006af2-d842-42f6-9191-74570d1de950, please check neutron logs for more information. [ 735.761713] env[63021]: ERROR nova.compute.manager [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] [ 735.762071] env[63021]: DEBUG nova.compute.utils [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Binding failed for port 9c006af2-d842-42f6-9191-74570d1de950, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 735.762755] env[63021]: DEBUG oslo_concurrency.lockutils [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.880s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.764386] env[63021]: INFO nova.compute.claims [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 735.768944] env[63021]: DEBUG nova.compute.manager [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Build of instance 03aec3ed-f338-41b5-a396-e3d693edbb8c was re-scheduled: Binding failed for port 9c006af2-d842-42f6-9191-74570d1de950, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 735.769399] env[63021]: DEBUG nova.compute.manager [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 735.769622] env[63021]: DEBUG oslo_concurrency.lockutils [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Acquiring lock "refresh_cache-03aec3ed-f338-41b5-a396-e3d693edbb8c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.771890] env[63021]: DEBUG oslo_concurrency.lockutils [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Acquired lock "refresh_cache-03aec3ed-f338-41b5-a396-e3d693edbb8c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.771890] env[63021]: DEBUG nova.network.neutron [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 735.799031] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lock "476e1e73-7050-4e53-bff8-d5d9f0982f1a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.935s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.042393] env[63021]: INFO nova.compute.manager [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 28d70e1d-da2a-4c9c-b297-86d026ce42c3] Took 1.04 seconds to deallocate network for instance. [ 736.291090] env[63021]: DEBUG nova.network.neutron [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 736.302688] env[63021]: DEBUG nova.compute.manager [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 736.373604] env[63021]: DEBUG nova.network.neutron [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.822905] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.875994] env[63021]: DEBUG oslo_concurrency.lockutils [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Releasing lock "refresh_cache-03aec3ed-f338-41b5-a396-e3d693edbb8c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.876316] env[63021]: DEBUG nova.compute.manager [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 736.876501] env[63021]: DEBUG nova.compute.manager [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 736.876668] env[63021]: DEBUG nova.network.neutron [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 736.891770] env[63021]: DEBUG nova.network.neutron [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 737.073269] env[63021]: INFO nova.scheduler.client.report [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Deleted allocations for instance 28d70e1d-da2a-4c9c-b297-86d026ce42c3 [ 737.093787] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abf42277-b5a5-400e-8045-9d5924f27a62 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.101022] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61f026b8-e8e9-4fb4-af87-2358f450320d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.141382] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-562e01b7-c5b1-4e81-91c5-a0daa9faa174 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.149561] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f67845db-e85a-45b3-87a9-e6ff0f581ef2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.163870] env[63021]: DEBUG nova.compute.provider_tree [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 737.394598] env[63021]: DEBUG nova.network.neutron [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.583858] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9e509fc9-d1e0-493d-bd1a-b6f0120ae25f tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lock "28d70e1d-da2a-4c9c-b297-86d026ce42c3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 139.653s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.666513] env[63021]: DEBUG nova.scheduler.client.report [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 737.898021] env[63021]: INFO nova.compute.manager [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] [instance: 03aec3ed-f338-41b5-a396-e3d693edbb8c] Took 1.02 seconds to deallocate network for instance. [ 738.086746] env[63021]: DEBUG nova.compute.manager [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 738.174455] env[63021]: DEBUG oslo_concurrency.lockutils [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.412s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.175143] env[63021]: DEBUG nova.compute.manager [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 738.177807] env[63021]: DEBUG oslo_concurrency.lockutils [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.586s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.609201] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.691614] env[63021]: DEBUG nova.compute.utils [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 738.694527] env[63021]: DEBUG nova.compute.manager [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 738.694617] env[63021]: DEBUG nova.network.neutron [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 274b001a-443f-463c-a657-652f64ba1de8] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 738.774182] env[63021]: DEBUG nova.policy [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '445719c275e5469fb60006288d6e02f3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ae1fd9195eb04d068d317938f79c54f0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 738.945303] env[63021]: INFO nova.scheduler.client.report [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Deleted allocations for instance 03aec3ed-f338-41b5-a396-e3d693edbb8c [ 739.065627] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c3916ad-c323-40d3-872f-3c5bacc7fa04 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.078465] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be5a310c-6f27-4671-a005-18ffad8e676c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.113937] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98b2677f-6ed9-4a70-94e2-cedd8bf731d0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.123022] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dfd9398-34eb-41dc-91f2-878331e82645 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.138133] env[63021]: DEBUG nova.compute.provider_tree [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 739.195387] env[63021]: DEBUG nova.compute.manager [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 739.243930] env[63021]: DEBUG nova.network.neutron [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Successfully created port: dda3fa9a-a2cf-4db2-a54f-f5cfe182e810 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 739.459351] env[63021]: DEBUG oslo_concurrency.lockutils [None req-eb4711e6-c8cd-4acd-9d83-1f72756bcfbf tempest-ServerActionsTestOtherB-978405642 tempest-ServerActionsTestOtherB-978405642-project-member] Lock "03aec3ed-f338-41b5-a396-e3d693edbb8c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.006s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.581269] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquiring lock "2209d709-7a5b-42e2-be93-b4fd436b6290" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.581506] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lock "2209d709-7a5b-42e2-be93-b4fd436b6290" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.618028] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquiring lock "36d26354-6127-4557-acb0-8743e27ff1ff" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.618494] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lock "36d26354-6127-4557-acb0-8743e27ff1ff" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.641670] env[63021]: DEBUG nova.scheduler.client.report [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 739.963123] env[63021]: DEBUG nova.compute.manager [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 740.150019] env[63021]: DEBUG oslo_concurrency.lockutils [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.969s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.150019] env[63021]: ERROR nova.compute.manager [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f20d29e3-a559-4952-9e91-3afed97b91b6, please check neutron logs for more information. [ 740.150019] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] Traceback (most recent call last): [ 740.150019] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 740.150019] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] self.driver.spawn(context, instance, image_meta, [ 740.150019] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 740.150019] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 740.150019] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 740.150019] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] vm_ref = self.build_virtual_machine(instance, [ 740.150439] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 740.150439] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] vif_infos = vmwarevif.get_vif_info(self._session, [ 740.150439] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 740.150439] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] for vif in network_info: [ 740.150439] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 740.150439] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] return self._sync_wrapper(fn, *args, **kwargs) [ 740.150439] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 740.150439] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] self.wait() [ 740.150439] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 740.150439] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] self[:] = self._gt.wait() [ 740.150439] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 740.150439] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] return self._exit_event.wait() [ 740.150439] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 740.150798] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] current.throw(*self._exc) [ 740.150798] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 740.150798] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] result = function(*args, **kwargs) [ 740.150798] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 740.150798] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] return func(*args, **kwargs) [ 740.150798] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 740.150798] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] raise e [ 740.150798] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 740.150798] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] nwinfo = self.network_api.allocate_for_instance( [ 740.150798] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 740.150798] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] created_port_ids = self._update_ports_for_instance( [ 740.150798] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 740.150798] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] with excutils.save_and_reraise_exception(): [ 740.151153] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 740.151153] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] self.force_reraise() [ 740.151153] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 740.151153] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] raise self.value [ 740.151153] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 740.151153] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] updated_port = self._update_port( [ 740.151153] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 740.151153] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] _ensure_no_port_binding_failure(port) [ 740.151153] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 740.151153] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] raise exception.PortBindingFailed(port_id=port['id']) [ 740.151153] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] nova.exception.PortBindingFailed: Binding failed for port f20d29e3-a559-4952-9e91-3afed97b91b6, please check neutron logs for more information. [ 740.151153] env[63021]: ERROR nova.compute.manager [instance: 239d0087-462d-445a-a131-a23d4d15505f] [ 740.151472] env[63021]: DEBUG nova.compute.utils [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Binding failed for port f20d29e3-a559-4952-9e91-3afed97b91b6, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 740.151472] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.833s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.157048] env[63021]: DEBUG nova.compute.manager [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Build of instance 239d0087-462d-445a-a131-a23d4d15505f was re-scheduled: Binding failed for port f20d29e3-a559-4952-9e91-3afed97b91b6, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 740.157048] env[63021]: DEBUG nova.compute.manager [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 740.157048] env[63021]: DEBUG oslo_concurrency.lockutils [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Acquiring lock "refresh_cache-239d0087-462d-445a-a131-a23d4d15505f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.157048] env[63021]: DEBUG oslo_concurrency.lockutils [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Acquired lock "refresh_cache-239d0087-462d-445a-a131-a23d4d15505f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.158407] env[63021]: DEBUG nova.network.neutron [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 740.186379] env[63021]: DEBUG nova.compute.manager [req-8ef51809-d904-4000-a1e4-dbc7d378dcd8 req-a8d5c4e4-c8c8-4929-bc7a-d8acb7943268 service nova] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Received event network-changed-dda3fa9a-a2cf-4db2-a54f-f5cfe182e810 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 740.186779] env[63021]: DEBUG nova.compute.manager [req-8ef51809-d904-4000-a1e4-dbc7d378dcd8 req-a8d5c4e4-c8c8-4929-bc7a-d8acb7943268 service nova] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Refreshing instance network info cache due to event network-changed-dda3fa9a-a2cf-4db2-a54f-f5cfe182e810. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 740.187292] env[63021]: DEBUG oslo_concurrency.lockutils [req-8ef51809-d904-4000-a1e4-dbc7d378dcd8 req-a8d5c4e4-c8c8-4929-bc7a-d8acb7943268 service nova] Acquiring lock "refresh_cache-274b001a-443f-463c-a657-652f64ba1de8" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.187568] env[63021]: DEBUG oslo_concurrency.lockutils [req-8ef51809-d904-4000-a1e4-dbc7d378dcd8 req-a8d5c4e4-c8c8-4929-bc7a-d8acb7943268 service nova] Acquired lock "refresh_cache-274b001a-443f-463c-a657-652f64ba1de8" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.189168] env[63021]: DEBUG nova.network.neutron [req-8ef51809-d904-4000-a1e4-dbc7d378dcd8 req-a8d5c4e4-c8c8-4929-bc7a-d8acb7943268 service nova] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Refreshing network info cache for port dda3fa9a-a2cf-4db2-a54f-f5cfe182e810 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 740.209513] env[63021]: DEBUG nova.compute.manager [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 740.237046] env[63021]: DEBUG nova.virt.hardware [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 740.237717] env[63021]: DEBUG nova.virt.hardware [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 740.237717] env[63021]: DEBUG nova.virt.hardware [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 740.237717] env[63021]: DEBUG nova.virt.hardware [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 740.237914] env[63021]: DEBUG nova.virt.hardware [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 740.237963] env[63021]: DEBUG nova.virt.hardware [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 740.238462] env[63021]: DEBUG nova.virt.hardware [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 740.238462] env[63021]: DEBUG nova.virt.hardware [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 740.238462] env[63021]: DEBUG nova.virt.hardware [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 740.238650] env[63021]: DEBUG nova.virt.hardware [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 740.238834] env[63021]: DEBUG nova.virt.hardware [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 740.239731] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-512ced6f-8075-4b32-b1fc-7bd4cc36c178 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.250991] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3292821c-a5a9-4a63-8117-e67f6e8c08a7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.464111] env[63021]: ERROR nova.compute.manager [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port dda3fa9a-a2cf-4db2-a54f-f5cfe182e810, please check neutron logs for more information. [ 740.464111] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 740.464111] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 740.464111] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 740.464111] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 740.464111] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 740.464111] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 740.464111] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 740.464111] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 740.464111] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 740.464111] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 740.464111] env[63021]: ERROR nova.compute.manager raise self.value [ 740.464111] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 740.464111] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 740.464111] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 740.464111] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 740.464652] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 740.464652] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 740.464652] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port dda3fa9a-a2cf-4db2-a54f-f5cfe182e810, please check neutron logs for more information. [ 740.464652] env[63021]: ERROR nova.compute.manager [ 740.464652] env[63021]: Traceback (most recent call last): [ 740.464652] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 740.464652] env[63021]: listener.cb(fileno) [ 740.464652] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 740.464652] env[63021]: result = function(*args, **kwargs) [ 740.464652] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 740.464652] env[63021]: return func(*args, **kwargs) [ 740.464652] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 740.464652] env[63021]: raise e [ 740.464652] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 740.464652] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 740.464652] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 740.464652] env[63021]: created_port_ids = self._update_ports_for_instance( [ 740.464652] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 740.464652] env[63021]: with excutils.save_and_reraise_exception(): [ 740.464652] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 740.464652] env[63021]: self.force_reraise() [ 740.464652] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 740.464652] env[63021]: raise self.value [ 740.464652] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 740.464652] env[63021]: updated_port = self._update_port( [ 740.464652] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 740.464652] env[63021]: _ensure_no_port_binding_failure(port) [ 740.464652] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 740.464652] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 740.465573] env[63021]: nova.exception.PortBindingFailed: Binding failed for port dda3fa9a-a2cf-4db2-a54f-f5cfe182e810, please check neutron logs for more information. [ 740.465573] env[63021]: Removing descriptor: 17 [ 740.465573] env[63021]: ERROR nova.compute.manager [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port dda3fa9a-a2cf-4db2-a54f-f5cfe182e810, please check neutron logs for more information. [ 740.465573] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] Traceback (most recent call last): [ 740.465573] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 740.465573] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] yield resources [ 740.465573] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 740.465573] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] self.driver.spawn(context, instance, image_meta, [ 740.465573] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 740.465573] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 740.465573] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 740.465573] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] vm_ref = self.build_virtual_machine(instance, [ 740.465963] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 740.465963] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] vif_infos = vmwarevif.get_vif_info(self._session, [ 740.465963] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 740.465963] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] for vif in network_info: [ 740.465963] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 740.465963] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] return self._sync_wrapper(fn, *args, **kwargs) [ 740.465963] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 740.465963] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] self.wait() [ 740.465963] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 740.465963] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] self[:] = self._gt.wait() [ 740.465963] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 740.465963] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] return self._exit_event.wait() [ 740.465963] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 740.466390] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] result = hub.switch() [ 740.466390] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 740.466390] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] return self.greenlet.switch() [ 740.466390] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 740.466390] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] result = function(*args, **kwargs) [ 740.466390] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 740.466390] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] return func(*args, **kwargs) [ 740.466390] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 740.466390] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] raise e [ 740.466390] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 740.466390] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] nwinfo = self.network_api.allocate_for_instance( [ 740.466390] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 740.466390] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] created_port_ids = self._update_ports_for_instance( [ 740.466804] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 740.466804] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] with excutils.save_and_reraise_exception(): [ 740.466804] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 740.466804] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] self.force_reraise() [ 740.466804] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 740.466804] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] raise self.value [ 740.466804] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 740.466804] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] updated_port = self._update_port( [ 740.466804] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 740.466804] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] _ensure_no_port_binding_failure(port) [ 740.466804] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 740.466804] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] raise exception.PortBindingFailed(port_id=port['id']) [ 740.467208] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] nova.exception.PortBindingFailed: Binding failed for port dda3fa9a-a2cf-4db2-a54f-f5cfe182e810, please check neutron logs for more information. [ 740.467208] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] [ 740.467208] env[63021]: INFO nova.compute.manager [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Terminating instance [ 740.471307] env[63021]: DEBUG oslo_concurrency.lockutils [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Acquiring lock "refresh_cache-274b001a-443f-463c-a657-652f64ba1de8" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.501890] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.680673] env[63021]: DEBUG nova.network.neutron [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.711333] env[63021]: DEBUG nova.network.neutron [req-8ef51809-d904-4000-a1e4-dbc7d378dcd8 req-a8d5c4e4-c8c8-4929-bc7a-d8acb7943268 service nova] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.776417] env[63021]: DEBUG nova.network.neutron [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.833737] env[63021]: DEBUG nova.network.neutron [req-8ef51809-d904-4000-a1e4-dbc7d378dcd8 req-a8d5c4e4-c8c8-4929-bc7a-d8acb7943268 service nova] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.074778] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cf6ea67-8628-443e-9724-eb3922672ad2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.083053] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-021a8f64-52dc-43a1-8aaa-bf62a122ff4c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.127029] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fca45278-012a-414c-a168-036741851716 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.137525] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21b0eee3-9209-49cf-9678-ab3c40a35ee8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.153908] env[63021]: DEBUG nova.compute.provider_tree [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 741.279484] env[63021]: DEBUG oslo_concurrency.lockutils [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Releasing lock "refresh_cache-239d0087-462d-445a-a131-a23d4d15505f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.279731] env[63021]: DEBUG nova.compute.manager [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 741.279915] env[63021]: DEBUG nova.compute.manager [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 741.280096] env[63021]: DEBUG nova.network.neutron [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] [instance: 239d0087-462d-445a-a131-a23d4d15505f] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 741.339366] env[63021]: DEBUG oslo_concurrency.lockutils [req-8ef51809-d904-4000-a1e4-dbc7d378dcd8 req-a8d5c4e4-c8c8-4929-bc7a-d8acb7943268 service nova] Releasing lock "refresh_cache-274b001a-443f-463c-a657-652f64ba1de8" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.339801] env[63021]: DEBUG oslo_concurrency.lockutils [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Acquired lock "refresh_cache-274b001a-443f-463c-a657-652f64ba1de8" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.339980] env[63021]: DEBUG nova.network.neutron [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 741.358839] env[63021]: DEBUG nova.network.neutron [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.658044] env[63021]: DEBUG nova.scheduler.client.report [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 741.813445] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Acquiring lock "2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.814575] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Lock "2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.862078] env[63021]: DEBUG nova.network.neutron [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.870138] env[63021]: DEBUG nova.network.neutron [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.973157] env[63021]: DEBUG nova.network.neutron [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.167120] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.014s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.167120] env[63021]: ERROR nova.compute.manager [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port cfe90de3-370b-4f29-8bbc-e29a70448987, please check neutron logs for more information. [ 742.167120] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] Traceback (most recent call last): [ 742.167120] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 742.167120] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] self.driver.spawn(context, instance, image_meta, [ 742.167120] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 742.167120] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 742.167120] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 742.167120] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] vm_ref = self.build_virtual_machine(instance, [ 742.167549] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 742.167549] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] vif_infos = vmwarevif.get_vif_info(self._session, [ 742.167549] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 742.167549] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] for vif in network_info: [ 742.167549] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 742.167549] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] return self._sync_wrapper(fn, *args, **kwargs) [ 742.167549] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 742.167549] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] self.wait() [ 742.167549] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 742.167549] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] self[:] = self._gt.wait() [ 742.167549] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 742.167549] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] return self._exit_event.wait() [ 742.167549] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 742.167973] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] result = hub.switch() [ 742.167973] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 742.167973] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] return self.greenlet.switch() [ 742.167973] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 742.167973] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] result = function(*args, **kwargs) [ 742.167973] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 742.167973] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] return func(*args, **kwargs) [ 742.167973] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 742.167973] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] raise e [ 742.167973] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 742.167973] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] nwinfo = self.network_api.allocate_for_instance( [ 742.167973] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 742.167973] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] created_port_ids = self._update_ports_for_instance( [ 742.168422] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 742.168422] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] with excutils.save_and_reraise_exception(): [ 742.168422] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 742.168422] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] self.force_reraise() [ 742.168422] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 742.168422] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] raise self.value [ 742.168422] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 742.168422] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] updated_port = self._update_port( [ 742.168422] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 742.168422] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] _ensure_no_port_binding_failure(port) [ 742.168422] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 742.168422] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] raise exception.PortBindingFailed(port_id=port['id']) [ 742.168798] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] nova.exception.PortBindingFailed: Binding failed for port cfe90de3-370b-4f29-8bbc-e29a70448987, please check neutron logs for more information. [ 742.168798] env[63021]: ERROR nova.compute.manager [instance: c2523d70-12dd-444b-8839-b97378ef687e] [ 742.168798] env[63021]: DEBUG nova.compute.utils [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Binding failed for port cfe90de3-370b-4f29-8bbc-e29a70448987, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 742.168798] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.517s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.174794] env[63021]: DEBUG nova.compute.manager [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Build of instance c2523d70-12dd-444b-8839-b97378ef687e was re-scheduled: Binding failed for port cfe90de3-370b-4f29-8bbc-e29a70448987, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 742.174794] env[63021]: DEBUG nova.compute.manager [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 742.174794] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Acquiring lock "refresh_cache-c2523d70-12dd-444b-8839-b97378ef687e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.174794] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Acquired lock "refresh_cache-c2523d70-12dd-444b-8839-b97378ef687e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.175111] env[63021]: DEBUG nova.network.neutron [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 742.215247] env[63021]: DEBUG nova.compute.manager [req-fa4ba81e-5fe9-4086-b7d9-bdfea4433603 req-0800fdc2-a94e-44d2-9a8f-5a305457828c service nova] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Received event network-vif-deleted-dda3fa9a-a2cf-4db2-a54f-f5cfe182e810 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 742.373168] env[63021]: INFO nova.compute.manager [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] [instance: 239d0087-462d-445a-a131-a23d4d15505f] Took 1.09 seconds to deallocate network for instance. [ 742.480819] env[63021]: DEBUG oslo_concurrency.lockutils [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Releasing lock "refresh_cache-274b001a-443f-463c-a657-652f64ba1de8" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.481279] env[63021]: DEBUG nova.compute.manager [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 742.481458] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 742.482021] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-df903b89-80b8-49cf-b654-0dffd78cb75e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.492053] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-692643c4-d67f-4971-8138-462456417baa {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.522135] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 274b001a-443f-463c-a657-652f64ba1de8 could not be found. [ 742.523174] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 742.523374] env[63021]: INFO nova.compute.manager [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Took 0.04 seconds to destroy the instance on the hypervisor. [ 742.523631] env[63021]: DEBUG oslo.service.loopingcall [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 742.523865] env[63021]: DEBUG nova.compute.manager [-] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 742.523958] env[63021]: DEBUG nova.network.neutron [-] [instance: 274b001a-443f-463c-a657-652f64ba1de8] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 742.540081] env[63021]: DEBUG nova.network.neutron [-] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 742.692435] env[63021]: DEBUG nova.network.neutron [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 742.815213] env[63021]: DEBUG nova.network.neutron [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.018845] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc298326-8358-4071-8c4d-98059389b86d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.026223] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c68a924-7331-4b50-b142-56a625d2663a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.055214] env[63021]: DEBUG nova.network.neutron [-] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.057324] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0c4636b-a7f9-4903-afbc-daef60e5151e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.065659] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c115791-a813-4eef-ac59-6059b47d9222 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.079606] env[63021]: DEBUG nova.compute.provider_tree [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 743.317895] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Releasing lock "refresh_cache-c2523d70-12dd-444b-8839-b97378ef687e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.317895] env[63021]: DEBUG nova.compute.manager [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 743.318086] env[63021]: DEBUG nova.compute.manager [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 743.318258] env[63021]: DEBUG nova.network.neutron [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] [instance: c2523d70-12dd-444b-8839-b97378ef687e] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 743.333685] env[63021]: DEBUG nova.network.neutron [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 743.403098] env[63021]: INFO nova.scheduler.client.report [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Deleted allocations for instance 239d0087-462d-445a-a131-a23d4d15505f [ 743.561237] env[63021]: INFO nova.compute.manager [-] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Took 1.04 seconds to deallocate network for instance. [ 743.564061] env[63021]: DEBUG nova.compute.claims [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 743.564061] env[63021]: DEBUG oslo_concurrency.lockutils [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.583086] env[63021]: DEBUG nova.scheduler.client.report [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 743.836105] env[63021]: DEBUG nova.network.neutron [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.912699] env[63021]: DEBUG oslo_concurrency.lockutils [None req-71d0c682-d939-4d6b-9fc6-0fd20a610f8a tempest-FloatingIPsAssociationTestJSON-1647106745 tempest-FloatingIPsAssociationTestJSON-1647106745-project-member] Lock "239d0087-462d-445a-a131-a23d4d15505f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 143.301s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.004681] env[63021]: DEBUG oslo_concurrency.lockutils [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Acquiring lock "6f4e6245-c0d9-46b4-baf5-5aca19e66da5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.008048] env[63021]: DEBUG oslo_concurrency.lockutils [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Lock "6f4e6245-c0d9-46b4-baf5-5aca19e66da5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.088710] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.922s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.089787] env[63021]: ERROR nova.compute.manager [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f0ca34f2-e0b1-42ec-8269-6bb0525ea8c0, please check neutron logs for more information. [ 744.089787] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Traceback (most recent call last): [ 744.089787] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 744.089787] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] self.driver.spawn(context, instance, image_meta, [ 744.089787] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 744.089787] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] self._vmops.spawn(context, instance, image_meta, injected_files, [ 744.089787] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 744.089787] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] vm_ref = self.build_virtual_machine(instance, [ 744.089787] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 744.089787] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] vif_infos = vmwarevif.get_vif_info(self._session, [ 744.089787] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 744.090193] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] for vif in network_info: [ 744.090193] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 744.090193] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] return self._sync_wrapper(fn, *args, **kwargs) [ 744.090193] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 744.090193] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] self.wait() [ 744.090193] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 744.090193] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] self[:] = self._gt.wait() [ 744.090193] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 744.090193] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] return self._exit_event.wait() [ 744.090193] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 744.090193] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] current.throw(*self._exc) [ 744.090193] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 744.090193] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] result = function(*args, **kwargs) [ 744.090604] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 744.090604] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] return func(*args, **kwargs) [ 744.090604] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 744.090604] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] raise e [ 744.090604] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 744.090604] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] nwinfo = self.network_api.allocate_for_instance( [ 744.090604] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 744.090604] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] created_port_ids = self._update_ports_for_instance( [ 744.090604] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 744.090604] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] with excutils.save_and_reraise_exception(): [ 744.090604] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 744.090604] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] self.force_reraise() [ 744.090604] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 744.091033] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] raise self.value [ 744.091033] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 744.091033] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] updated_port = self._update_port( [ 744.091033] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 744.091033] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] _ensure_no_port_binding_failure(port) [ 744.091033] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 744.091033] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] raise exception.PortBindingFailed(port_id=port['id']) [ 744.091033] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] nova.exception.PortBindingFailed: Binding failed for port f0ca34f2-e0b1-42ec-8269-6bb0525ea8c0, please check neutron logs for more information. [ 744.091033] env[63021]: ERROR nova.compute.manager [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] [ 744.091033] env[63021]: DEBUG nova.compute.utils [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Binding failed for port f0ca34f2-e0b1-42ec-8269-6bb0525ea8c0, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 744.092698] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.153s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.094115] env[63021]: INFO nova.compute.claims [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 744.096654] env[63021]: DEBUG nova.compute.manager [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Build of instance 26896531-07eb-4e44-8dd8-e5237bfd7c39 was re-scheduled: Binding failed for port f0ca34f2-e0b1-42ec-8269-6bb0525ea8c0, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 744.097089] env[63021]: DEBUG nova.compute.manager [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 744.097311] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Acquiring lock "refresh_cache-26896531-07eb-4e44-8dd8-e5237bfd7c39" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.097453] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Acquired lock "refresh_cache-26896531-07eb-4e44-8dd8-e5237bfd7c39" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.097607] env[63021]: DEBUG nova.network.neutron [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 744.338433] env[63021]: INFO nova.compute.manager [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] [instance: c2523d70-12dd-444b-8839-b97378ef687e] Took 1.02 seconds to deallocate network for instance. [ 744.418399] env[63021]: DEBUG nova.compute.manager [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] [instance: f77da305-9cc9-415d-a808-c2136212c456] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 744.501156] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 744.501156] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 744.626109] env[63021]: DEBUG nova.network.neutron [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 744.711873] env[63021]: DEBUG nova.network.neutron [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.954406] env[63021]: DEBUG oslo_concurrency.lockutils [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.006872] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 745.007087] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Starting heal instance info cache {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 745.007168] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Rebuilding the list of instances to heal {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 745.215782] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Releasing lock "refresh_cache-26896531-07eb-4e44-8dd8-e5237bfd7c39" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.216010] env[63021]: DEBUG nova.compute.manager [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 745.216196] env[63021]: DEBUG nova.compute.manager [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 745.216359] env[63021]: DEBUG nova.network.neutron [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 745.236784] env[63021]: DEBUG nova.network.neutron [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 745.373011] env[63021]: INFO nova.scheduler.client.report [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Deleted allocations for instance c2523d70-12dd-444b-8839-b97378ef687e [ 745.474708] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-374bc951-1e1b-43f8-8580-134a73522793 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.483897] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a42a88c8-54f4-4a32-ae59-302c91599cf7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.521010] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Skipping network cache update for instance because it is Building. {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 745.521332] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Skipping network cache update for instance because it is Building. {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 745.521468] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Didn't find any instances for network info cache update. {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 745.522088] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 745.523199] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e23c7bf-50b2-439d-a799-4fdd80801910 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.526976] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 745.527221] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 745.527972] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 745.528206] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 745.528426] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 745.528616] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63021) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 745.528824] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 745.534679] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5557042d-13cb-4e3f-bd93-709869989172 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.558466] env[63021]: DEBUG nova.compute.provider_tree [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 745.739285] env[63021]: DEBUG nova.network.neutron [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.887253] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0940838c-737f-4220-a9b4-f9e5bfe8f86b tempest-InstanceActionsTestJSON-1805077702 tempest-InstanceActionsTestJSON-1805077702-project-member] Lock "c2523d70-12dd-444b-8839-b97378ef687e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.412s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.034800] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.063239] env[63021]: DEBUG nova.scheduler.client.report [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 746.242963] env[63021]: INFO nova.compute.manager [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 26896531-07eb-4e44-8dd8-e5237bfd7c39] Took 1.03 seconds to deallocate network for instance. [ 746.393025] env[63021]: DEBUG nova.compute.manager [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 746.571020] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.476s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.571020] env[63021]: DEBUG nova.compute.manager [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 746.572387] env[63021]: DEBUG oslo_concurrency.lockutils [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.721s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.575214] env[63021]: INFO nova.compute.claims [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 746.926527] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.082153] env[63021]: DEBUG nova.compute.utils [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 747.083635] env[63021]: DEBUG nova.compute.manager [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 747.085235] env[63021]: DEBUG nova.network.neutron [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 747.160863] env[63021]: DEBUG nova.policy [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '445719c275e5469fb60006288d6e02f3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ae1fd9195eb04d068d317938f79c54f0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 747.278249] env[63021]: INFO nova.scheduler.client.report [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Deleted allocations for instance 26896531-07eb-4e44-8dd8-e5237bfd7c39 [ 747.584624] env[63021]: DEBUG nova.compute.manager [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 747.649687] env[63021]: DEBUG nova.network.neutron [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Successfully created port: d88f0cc0-bc74-4b9e-842a-138776ad2023 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 747.787648] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6b3b6114-f65f-47ae-ae0d-581d2c13a30d tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Lock "26896531-07eb-4e44-8dd8-e5237bfd7c39" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.535s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.991599] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a8dd296-ea39-4674-b69f-f7fef2d29d72 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.000741] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de663d2f-4fa0-45bf-9537-b201509177cd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.032767] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f99dcd0-11c1-4571-8e3d-7029a81b5ab0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.041807] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f43d1045-e5ec-4ec3-9b35-2f4849305405 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.059051] env[63021]: DEBUG nova.compute.provider_tree [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 748.289328] env[63021]: DEBUG nova.compute.manager [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 748.563829] env[63021]: DEBUG nova.scheduler.client.report [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 748.592842] env[63021]: DEBUG nova.compute.manager [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 748.620536] env[63021]: DEBUG nova.virt.hardware [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 748.620774] env[63021]: DEBUG nova.virt.hardware [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 748.620927] env[63021]: DEBUG nova.virt.hardware [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 748.621285] env[63021]: DEBUG nova.virt.hardware [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 748.621393] env[63021]: DEBUG nova.virt.hardware [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 748.621583] env[63021]: DEBUG nova.virt.hardware [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 748.621826] env[63021]: DEBUG nova.virt.hardware [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 748.622022] env[63021]: DEBUG nova.virt.hardware [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 748.622272] env[63021]: DEBUG nova.virt.hardware [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 748.622421] env[63021]: DEBUG nova.virt.hardware [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 748.622667] env[63021]: DEBUG nova.virt.hardware [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 748.623860] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27c7e54b-0378-47cf-b57a-ed2d3de0bd8b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.632108] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-875670c1-10d5-45cf-b66a-e110d5b54c91 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.820783] env[63021]: DEBUG oslo_concurrency.lockutils [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.986816] env[63021]: DEBUG nova.compute.manager [req-9b52ffa1-aa0f-4b97-83bf-41ff6420e8a9 req-53919f5d-92ed-4a75-af7c-70b8b6ada1e0 service nova] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Received event network-changed-d88f0cc0-bc74-4b9e-842a-138776ad2023 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 748.987121] env[63021]: DEBUG nova.compute.manager [req-9b52ffa1-aa0f-4b97-83bf-41ff6420e8a9 req-53919f5d-92ed-4a75-af7c-70b8b6ada1e0 service nova] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Refreshing instance network info cache due to event network-changed-d88f0cc0-bc74-4b9e-842a-138776ad2023. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 748.987227] env[63021]: DEBUG oslo_concurrency.lockutils [req-9b52ffa1-aa0f-4b97-83bf-41ff6420e8a9 req-53919f5d-92ed-4a75-af7c-70b8b6ada1e0 service nova] Acquiring lock "refresh_cache-3660a926-7460-41f2-9ee4-2a5072701aea" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.987362] env[63021]: DEBUG oslo_concurrency.lockutils [req-9b52ffa1-aa0f-4b97-83bf-41ff6420e8a9 req-53919f5d-92ed-4a75-af7c-70b8b6ada1e0 service nova] Acquired lock "refresh_cache-3660a926-7460-41f2-9ee4-2a5072701aea" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.987587] env[63021]: DEBUG nova.network.neutron [req-9b52ffa1-aa0f-4b97-83bf-41ff6420e8a9 req-53919f5d-92ed-4a75-af7c-70b8b6ada1e0 service nova] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Refreshing network info cache for port d88f0cc0-bc74-4b9e-842a-138776ad2023 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 749.071156] env[63021]: DEBUG oslo_concurrency.lockutils [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.499s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.071638] env[63021]: DEBUG nova.compute.manager [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 749.074634] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.327s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.078543] env[63021]: INFO nova.compute.claims [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 749.199259] env[63021]: ERROR nova.compute.manager [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d88f0cc0-bc74-4b9e-842a-138776ad2023, please check neutron logs for more information. [ 749.199259] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 749.199259] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.199259] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 749.199259] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 749.199259] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 749.199259] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 749.199259] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 749.199259] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.199259] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 749.199259] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.199259] env[63021]: ERROR nova.compute.manager raise self.value [ 749.199259] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 749.199259] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 749.199259] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.199259] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 749.199790] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.199790] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 749.199790] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d88f0cc0-bc74-4b9e-842a-138776ad2023, please check neutron logs for more information. [ 749.199790] env[63021]: ERROR nova.compute.manager [ 749.199790] env[63021]: Traceback (most recent call last): [ 749.199790] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 749.199790] env[63021]: listener.cb(fileno) [ 749.199790] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 749.199790] env[63021]: result = function(*args, **kwargs) [ 749.199790] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 749.199790] env[63021]: return func(*args, **kwargs) [ 749.199790] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 749.199790] env[63021]: raise e [ 749.199790] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.199790] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 749.199790] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 749.199790] env[63021]: created_port_ids = self._update_ports_for_instance( [ 749.199790] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 749.199790] env[63021]: with excutils.save_and_reraise_exception(): [ 749.199790] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.199790] env[63021]: self.force_reraise() [ 749.199790] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.199790] env[63021]: raise self.value [ 749.199790] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 749.199790] env[63021]: updated_port = self._update_port( [ 749.199790] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.199790] env[63021]: _ensure_no_port_binding_failure(port) [ 749.199790] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.199790] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 749.200704] env[63021]: nova.exception.PortBindingFailed: Binding failed for port d88f0cc0-bc74-4b9e-842a-138776ad2023, please check neutron logs for more information. [ 749.200704] env[63021]: Removing descriptor: 18 [ 749.200704] env[63021]: ERROR nova.compute.manager [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d88f0cc0-bc74-4b9e-842a-138776ad2023, please check neutron logs for more information. [ 749.200704] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Traceback (most recent call last): [ 749.200704] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 749.200704] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] yield resources [ 749.200704] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 749.200704] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] self.driver.spawn(context, instance, image_meta, [ 749.200704] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 749.200704] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] self._vmops.spawn(context, instance, image_meta, injected_files, [ 749.200704] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 749.200704] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] vm_ref = self.build_virtual_machine(instance, [ 749.201056] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 749.201056] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] vif_infos = vmwarevif.get_vif_info(self._session, [ 749.201056] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 749.201056] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] for vif in network_info: [ 749.201056] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 749.201056] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] return self._sync_wrapper(fn, *args, **kwargs) [ 749.201056] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 749.201056] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] self.wait() [ 749.201056] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 749.201056] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] self[:] = self._gt.wait() [ 749.201056] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 749.201056] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] return self._exit_event.wait() [ 749.201056] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 749.201419] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] result = hub.switch() [ 749.201419] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 749.201419] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] return self.greenlet.switch() [ 749.201419] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 749.201419] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] result = function(*args, **kwargs) [ 749.201419] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 749.201419] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] return func(*args, **kwargs) [ 749.201419] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 749.201419] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] raise e [ 749.201419] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.201419] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] nwinfo = self.network_api.allocate_for_instance( [ 749.201419] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 749.201419] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] created_port_ids = self._update_ports_for_instance( [ 749.201786] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 749.201786] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] with excutils.save_and_reraise_exception(): [ 749.201786] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.201786] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] self.force_reraise() [ 749.201786] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.201786] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] raise self.value [ 749.201786] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 749.201786] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] updated_port = self._update_port( [ 749.201786] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.201786] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] _ensure_no_port_binding_failure(port) [ 749.201786] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.201786] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] raise exception.PortBindingFailed(port_id=port['id']) [ 749.202121] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] nova.exception.PortBindingFailed: Binding failed for port d88f0cc0-bc74-4b9e-842a-138776ad2023, please check neutron logs for more information. [ 749.202121] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] [ 749.202121] env[63021]: INFO nova.compute.manager [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Terminating instance [ 749.203509] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Acquiring lock "refresh_cache-3660a926-7460-41f2-9ee4-2a5072701aea" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.507215] env[63021]: DEBUG nova.network.neutron [req-9b52ffa1-aa0f-4b97-83bf-41ff6420e8a9 req-53919f5d-92ed-4a75-af7c-70b8b6ada1e0 service nova] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 749.580619] env[63021]: DEBUG nova.compute.utils [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 749.585113] env[63021]: DEBUG nova.compute.manager [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 749.589232] env[63021]: DEBUG nova.network.neutron [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 749.596162] env[63021]: DEBUG nova.network.neutron [req-9b52ffa1-aa0f-4b97-83bf-41ff6420e8a9 req-53919f5d-92ed-4a75-af7c-70b8b6ada1e0 service nova] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.689332] env[63021]: DEBUG nova.policy [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ee5c9d43339f4344a64cc24d68194653', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2ce7715ac3524d889e03c799cef7bb2c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 750.085914] env[63021]: DEBUG nova.compute.manager [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 750.097825] env[63021]: DEBUG oslo_concurrency.lockutils [req-9b52ffa1-aa0f-4b97-83bf-41ff6420e8a9 req-53919f5d-92ed-4a75-af7c-70b8b6ada1e0 service nova] Releasing lock "refresh_cache-3660a926-7460-41f2-9ee4-2a5072701aea" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.100374] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Acquired lock "refresh_cache-3660a926-7460-41f2-9ee4-2a5072701aea" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.100579] env[63021]: DEBUG nova.network.neutron [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 750.391170] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ff52d26-f282-40cf-a648-9b0d6eb45d76 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.399304] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50e7ea9b-614e-4581-8173-a32c5537d376 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.430452] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-811babc2-47ef-4fd2-896c-1eb6487ac747 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.437893] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ce6076c-268e-41cb-a915-1ea8b41b1374 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.452071] env[63021]: DEBUG nova.compute.provider_tree [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 750.500392] env[63021]: DEBUG nova.network.neutron [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Successfully created port: 09f2e8a1-f441-4cee-896f-7b33e52c6b5a {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 750.660317] env[63021]: DEBUG nova.network.neutron [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.779597] env[63021]: DEBUG nova.network.neutron [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.955808] env[63021]: DEBUG nova.scheduler.client.report [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 751.097313] env[63021]: DEBUG nova.compute.manager [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 751.120801] env[63021]: DEBUG nova.compute.manager [req-b44fcc25-db94-4c33-b23f-dd27f4766ed9 req-ea620500-5701-494d-88d3-cfeec1916f48 service nova] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Received event network-vif-deleted-d88f0cc0-bc74-4b9e-842a-138776ad2023 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 751.133215] env[63021]: DEBUG nova.virt.hardware [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 751.133452] env[63021]: DEBUG nova.virt.hardware [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 751.133605] env[63021]: DEBUG nova.virt.hardware [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 751.133813] env[63021]: DEBUG nova.virt.hardware [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 751.133961] env[63021]: DEBUG nova.virt.hardware [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 751.134116] env[63021]: DEBUG nova.virt.hardware [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 751.134322] env[63021]: DEBUG nova.virt.hardware [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 751.134476] env[63021]: DEBUG nova.virt.hardware [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 751.134635] env[63021]: DEBUG nova.virt.hardware [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 751.134795] env[63021]: DEBUG nova.virt.hardware [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 751.134960] env[63021]: DEBUG nova.virt.hardware [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 751.136044] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bc8fd09-06d6-42df-8f6f-426c5160604e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.144777] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96a1163b-bcc9-4fdb-8688-54a502525ce5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.282590] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Releasing lock "refresh_cache-3660a926-7460-41f2-9ee4-2a5072701aea" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.283069] env[63021]: DEBUG nova.compute.manager [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 751.283273] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 751.283575] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7d8921c3-6254-4b7f-be4b-d22a872641b7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.292817] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e82ac38-9c78-41a4-9b54-a2ace8f3998c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.317264] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Acquiring lock "70f67905-7c60-433d-9ebc-d66fa44eb36e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.317583] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Lock "70f67905-7c60-433d-9ebc-d66fa44eb36e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.318117] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3660a926-7460-41f2-9ee4-2a5072701aea could not be found. [ 751.318365] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 751.318592] env[63021]: INFO nova.compute.manager [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Took 0.04 seconds to destroy the instance on the hypervisor. [ 751.318858] env[63021]: DEBUG oslo.service.loopingcall [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 751.319113] env[63021]: DEBUG nova.compute.manager [-] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 751.319236] env[63021]: DEBUG nova.network.neutron [-] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 751.335334] env[63021]: DEBUG nova.network.neutron [-] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.461808] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.387s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.462365] env[63021]: DEBUG nova.compute.manager [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 751.465058] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.642s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.471823] env[63021]: INFO nova.compute.claims [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 751.517763] env[63021]: DEBUG nova.compute.manager [req-83583512-706e-48b5-a69b-fa192dfe7eb4 req-43949882-532f-4285-ac9b-0b4059275507 service nova] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Received event network-changed-09f2e8a1-f441-4cee-896f-7b33e52c6b5a {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 751.517968] env[63021]: DEBUG nova.compute.manager [req-83583512-706e-48b5-a69b-fa192dfe7eb4 req-43949882-532f-4285-ac9b-0b4059275507 service nova] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Refreshing instance network info cache due to event network-changed-09f2e8a1-f441-4cee-896f-7b33e52c6b5a. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 751.518642] env[63021]: DEBUG oslo_concurrency.lockutils [req-83583512-706e-48b5-a69b-fa192dfe7eb4 req-43949882-532f-4285-ac9b-0b4059275507 service nova] Acquiring lock "refresh_cache-bc48ac80-36f6-4c97-9f95-29aed2efeae6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.518802] env[63021]: DEBUG oslo_concurrency.lockutils [req-83583512-706e-48b5-a69b-fa192dfe7eb4 req-43949882-532f-4285-ac9b-0b4059275507 service nova] Acquired lock "refresh_cache-bc48ac80-36f6-4c97-9f95-29aed2efeae6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.518967] env[63021]: DEBUG nova.network.neutron [req-83583512-706e-48b5-a69b-fa192dfe7eb4 req-43949882-532f-4285-ac9b-0b4059275507 service nova] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Refreshing network info cache for port 09f2e8a1-f441-4cee-896f-7b33e52c6b5a {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 751.783335] env[63021]: ERROR nova.compute.manager [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 09f2e8a1-f441-4cee-896f-7b33e52c6b5a, please check neutron logs for more information. [ 751.783335] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 751.783335] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 751.783335] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 751.783335] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 751.783335] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 751.783335] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 751.783335] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 751.783335] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 751.783335] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 751.783335] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 751.783335] env[63021]: ERROR nova.compute.manager raise self.value [ 751.783335] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 751.783335] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 751.783335] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 751.783335] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 751.784071] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 751.784071] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 751.784071] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 09f2e8a1-f441-4cee-896f-7b33e52c6b5a, please check neutron logs for more information. [ 751.784071] env[63021]: ERROR nova.compute.manager [ 751.784071] env[63021]: Traceback (most recent call last): [ 751.784071] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 751.784071] env[63021]: listener.cb(fileno) [ 751.784071] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 751.784071] env[63021]: result = function(*args, **kwargs) [ 751.784071] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 751.784071] env[63021]: return func(*args, **kwargs) [ 751.784071] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 751.784071] env[63021]: raise e [ 751.784071] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 751.784071] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 751.784071] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 751.784071] env[63021]: created_port_ids = self._update_ports_for_instance( [ 751.784071] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 751.784071] env[63021]: with excutils.save_and_reraise_exception(): [ 751.784071] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 751.784071] env[63021]: self.force_reraise() [ 751.784071] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 751.784071] env[63021]: raise self.value [ 751.784071] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 751.784071] env[63021]: updated_port = self._update_port( [ 751.784071] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 751.784071] env[63021]: _ensure_no_port_binding_failure(port) [ 751.784071] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 751.784071] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 751.785020] env[63021]: nova.exception.PortBindingFailed: Binding failed for port 09f2e8a1-f441-4cee-896f-7b33e52c6b5a, please check neutron logs for more information. [ 751.785020] env[63021]: Removing descriptor: 17 [ 751.785020] env[63021]: ERROR nova.compute.manager [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 09f2e8a1-f441-4cee-896f-7b33e52c6b5a, please check neutron logs for more information. [ 751.785020] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Traceback (most recent call last): [ 751.785020] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 751.785020] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] yield resources [ 751.785020] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 751.785020] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] self.driver.spawn(context, instance, image_meta, [ 751.785020] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 751.785020] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 751.785020] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 751.785020] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] vm_ref = self.build_virtual_machine(instance, [ 751.785412] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 751.785412] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] vif_infos = vmwarevif.get_vif_info(self._session, [ 751.785412] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 751.785412] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] for vif in network_info: [ 751.785412] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 751.785412] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] return self._sync_wrapper(fn, *args, **kwargs) [ 751.785412] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 751.785412] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] self.wait() [ 751.785412] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 751.785412] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] self[:] = self._gt.wait() [ 751.785412] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 751.785412] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] return self._exit_event.wait() [ 751.785412] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 751.785827] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] result = hub.switch() [ 751.785827] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 751.785827] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] return self.greenlet.switch() [ 751.785827] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 751.785827] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] result = function(*args, **kwargs) [ 751.785827] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 751.785827] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] return func(*args, **kwargs) [ 751.785827] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 751.785827] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] raise e [ 751.785827] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 751.785827] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] nwinfo = self.network_api.allocate_for_instance( [ 751.785827] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 751.785827] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] created_port_ids = self._update_ports_for_instance( [ 751.786272] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 751.786272] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] with excutils.save_and_reraise_exception(): [ 751.786272] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 751.786272] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] self.force_reraise() [ 751.786272] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 751.786272] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] raise self.value [ 751.786272] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 751.786272] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] updated_port = self._update_port( [ 751.786272] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 751.786272] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] _ensure_no_port_binding_failure(port) [ 751.786272] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 751.786272] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] raise exception.PortBindingFailed(port_id=port['id']) [ 751.786864] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] nova.exception.PortBindingFailed: Binding failed for port 09f2e8a1-f441-4cee-896f-7b33e52c6b5a, please check neutron logs for more information. [ 751.786864] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] [ 751.786864] env[63021]: INFO nova.compute.manager [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Terminating instance [ 751.787772] env[63021]: DEBUG oslo_concurrency.lockutils [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Acquiring lock "refresh_cache-bc48ac80-36f6-4c97-9f95-29aed2efeae6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.838180] env[63021]: DEBUG nova.network.neutron [-] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.979036] env[63021]: DEBUG nova.compute.utils [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 751.987044] env[63021]: DEBUG nova.compute.manager [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 751.987044] env[63021]: DEBUG nova.network.neutron [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 752.046315] env[63021]: DEBUG nova.network.neutron [req-83583512-706e-48b5-a69b-fa192dfe7eb4 req-43949882-532f-4285-ac9b-0b4059275507 service nova] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 752.077529] env[63021]: DEBUG nova.policy [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2642402689b04f89946802ac3b3e6287', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '805af233efa147b3982fb794cd9e6f73', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 752.135233] env[63021]: DEBUG nova.network.neutron [req-83583512-706e-48b5-a69b-fa192dfe7eb4 req-43949882-532f-4285-ac9b-0b4059275507 service nova] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.341687] env[63021]: INFO nova.compute.manager [-] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Took 1.02 seconds to deallocate network for instance. [ 752.344046] env[63021]: DEBUG nova.compute.claims [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 752.344234] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.410306] env[63021]: DEBUG nova.network.neutron [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Successfully created port: 5cea81bb-fd5c-430e-b8f1-5013d52a6802 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 752.488865] env[63021]: DEBUG nova.compute.manager [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 752.535673] env[63021]: DEBUG oslo_concurrency.lockutils [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Acquiring lock "7787f6bb-e855-476f-9146-2a648932429d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.535913] env[63021]: DEBUG oslo_concurrency.lockutils [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Lock "7787f6bb-e855-476f-9146-2a648932429d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.639023] env[63021]: DEBUG oslo_concurrency.lockutils [req-83583512-706e-48b5-a69b-fa192dfe7eb4 req-43949882-532f-4285-ac9b-0b4059275507 service nova] Releasing lock "refresh_cache-bc48ac80-36f6-4c97-9f95-29aed2efeae6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.639023] env[63021]: DEBUG oslo_concurrency.lockutils [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Acquired lock "refresh_cache-bc48ac80-36f6-4c97-9f95-29aed2efeae6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.639023] env[63021]: DEBUG nova.network.neutron [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 752.842175] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f29d6244-5d66-431a-b081-ace088e7347f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.849198] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0a794a7-85d8-4abd-bc19-1dea8b0145b9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.885295] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a63ef007-9c9e-48ce-a092-b6f13d3587dd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.893237] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f090aa6-b868-4e87-9e08-3d832e3b702c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.910358] env[63021]: DEBUG nova.compute.provider_tree [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 753.164455] env[63021]: DEBUG nova.network.neutron [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 753.242756] env[63021]: DEBUG nova.network.neutron [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.413384] env[63021]: DEBUG nova.scheduler.client.report [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 753.496389] env[63021]: ERROR nova.compute.manager [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5cea81bb-fd5c-430e-b8f1-5013d52a6802, please check neutron logs for more information. [ 753.496389] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 753.496389] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 753.496389] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 753.496389] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 753.496389] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 753.496389] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 753.496389] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 753.496389] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.496389] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 753.496389] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.496389] env[63021]: ERROR nova.compute.manager raise self.value [ 753.496389] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 753.496389] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 753.496389] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.496389] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 753.496920] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.496920] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 753.496920] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5cea81bb-fd5c-430e-b8f1-5013d52a6802, please check neutron logs for more information. [ 753.496920] env[63021]: ERROR nova.compute.manager [ 753.496920] env[63021]: Traceback (most recent call last): [ 753.496920] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 753.496920] env[63021]: listener.cb(fileno) [ 753.496920] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 753.496920] env[63021]: result = function(*args, **kwargs) [ 753.496920] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 753.496920] env[63021]: return func(*args, **kwargs) [ 753.496920] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 753.496920] env[63021]: raise e [ 753.496920] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 753.496920] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 753.496920] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 753.496920] env[63021]: created_port_ids = self._update_ports_for_instance( [ 753.496920] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 753.496920] env[63021]: with excutils.save_and_reraise_exception(): [ 753.496920] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.496920] env[63021]: self.force_reraise() [ 753.496920] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.496920] env[63021]: raise self.value [ 753.496920] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 753.496920] env[63021]: updated_port = self._update_port( [ 753.496920] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.496920] env[63021]: _ensure_no_port_binding_failure(port) [ 753.496920] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.496920] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 753.497808] env[63021]: nova.exception.PortBindingFailed: Binding failed for port 5cea81bb-fd5c-430e-b8f1-5013d52a6802, please check neutron logs for more information. [ 753.497808] env[63021]: Removing descriptor: 17 [ 753.501619] env[63021]: DEBUG nova.compute.manager [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 753.527386] env[63021]: DEBUG nova.virt.hardware [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 753.527648] env[63021]: DEBUG nova.virt.hardware [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 753.527802] env[63021]: DEBUG nova.virt.hardware [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 753.527978] env[63021]: DEBUG nova.virt.hardware [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 753.528373] env[63021]: DEBUG nova.virt.hardware [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 753.528373] env[63021]: DEBUG nova.virt.hardware [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 753.528490] env[63021]: DEBUG nova.virt.hardware [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 753.528601] env[63021]: DEBUG nova.virt.hardware [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 753.528759] env[63021]: DEBUG nova.virt.hardware [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 753.528912] env[63021]: DEBUG nova.virt.hardware [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 753.529084] env[63021]: DEBUG nova.virt.hardware [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 753.529971] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc26d88b-2af7-41be-97fc-ae70ff1b01d3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.537822] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-735758bc-1f24-4ebc-aec3-45354ba45563 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.551471] env[63021]: ERROR nova.compute.manager [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5cea81bb-fd5c-430e-b8f1-5013d52a6802, please check neutron logs for more information. [ 753.551471] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Traceback (most recent call last): [ 753.551471] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 753.551471] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] yield resources [ 753.551471] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 753.551471] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] self.driver.spawn(context, instance, image_meta, [ 753.551471] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 753.551471] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] self._vmops.spawn(context, instance, image_meta, injected_files, [ 753.551471] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 753.551471] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] vm_ref = self.build_virtual_machine(instance, [ 753.551471] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 753.551811] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] vif_infos = vmwarevif.get_vif_info(self._session, [ 753.551811] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 753.551811] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] for vif in network_info: [ 753.551811] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 753.551811] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] return self._sync_wrapper(fn, *args, **kwargs) [ 753.551811] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 753.551811] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] self.wait() [ 753.551811] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 753.551811] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] self[:] = self._gt.wait() [ 753.551811] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 753.551811] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] return self._exit_event.wait() [ 753.551811] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 753.551811] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] current.throw(*self._exc) [ 753.552419] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 753.552419] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] result = function(*args, **kwargs) [ 753.552419] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 753.552419] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] return func(*args, **kwargs) [ 753.552419] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 753.552419] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] raise e [ 753.552419] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 753.552419] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] nwinfo = self.network_api.allocate_for_instance( [ 753.552419] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 753.552419] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] created_port_ids = self._update_ports_for_instance( [ 753.552419] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 753.552419] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] with excutils.save_and_reraise_exception(): [ 753.552419] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.553322] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] self.force_reraise() [ 753.553322] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.553322] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] raise self.value [ 753.553322] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 753.553322] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] updated_port = self._update_port( [ 753.553322] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.553322] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] _ensure_no_port_binding_failure(port) [ 753.553322] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.553322] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] raise exception.PortBindingFailed(port_id=port['id']) [ 753.553322] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] nova.exception.PortBindingFailed: Binding failed for port 5cea81bb-fd5c-430e-b8f1-5013d52a6802, please check neutron logs for more information. [ 753.553322] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] [ 753.553322] env[63021]: INFO nova.compute.manager [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Terminating instance [ 753.555132] env[63021]: DEBUG nova.compute.manager [req-adae83f7-da48-4003-a03c-08b535a71e54 req-bbe2c37e-1fde-4e2a-8032-7f1e004974f0 service nova] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Received event network-vif-deleted-09f2e8a1-f441-4cee-896f-7b33e52c6b5a {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 753.555328] env[63021]: DEBUG nova.compute.manager [req-adae83f7-da48-4003-a03c-08b535a71e54 req-bbe2c37e-1fde-4e2a-8032-7f1e004974f0 service nova] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Received event network-changed-5cea81bb-fd5c-430e-b8f1-5013d52a6802 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 753.555485] env[63021]: DEBUG nova.compute.manager [req-adae83f7-da48-4003-a03c-08b535a71e54 req-bbe2c37e-1fde-4e2a-8032-7f1e004974f0 service nova] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Refreshing instance network info cache due to event network-changed-5cea81bb-fd5c-430e-b8f1-5013d52a6802. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 753.555684] env[63021]: DEBUG oslo_concurrency.lockutils [req-adae83f7-da48-4003-a03c-08b535a71e54 req-bbe2c37e-1fde-4e2a-8032-7f1e004974f0 service nova] Acquiring lock "refresh_cache-907540da-a701-477a-9b5c-9942b5d2d987" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.555821] env[63021]: DEBUG oslo_concurrency.lockutils [req-adae83f7-da48-4003-a03c-08b535a71e54 req-bbe2c37e-1fde-4e2a-8032-7f1e004974f0 service nova] Acquired lock "refresh_cache-907540da-a701-477a-9b5c-9942b5d2d987" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.555970] env[63021]: DEBUG nova.network.neutron [req-adae83f7-da48-4003-a03c-08b535a71e54 req-bbe2c37e-1fde-4e2a-8032-7f1e004974f0 service nova] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Refreshing network info cache for port 5cea81bb-fd5c-430e-b8f1-5013d52a6802 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 753.558186] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "refresh_cache-907540da-a701-477a-9b5c-9942b5d2d987" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.745747] env[63021]: DEBUG oslo_concurrency.lockutils [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Releasing lock "refresh_cache-bc48ac80-36f6-4c97-9f95-29aed2efeae6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.746277] env[63021]: DEBUG nova.compute.manager [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 753.746498] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 753.746801] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-db00377e-97b9-4cda-8bdf-0d29c3e5d2d0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.756546] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4c478df-3493-4d1f-af8a-4c37ce2afa21 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.777107] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bc48ac80-36f6-4c97-9f95-29aed2efeae6 could not be found. [ 753.777288] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 753.777468] env[63021]: INFO nova.compute.manager [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Took 0.03 seconds to destroy the instance on the hypervisor. [ 753.777701] env[63021]: DEBUG oslo.service.loopingcall [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 753.777916] env[63021]: DEBUG nova.compute.manager [-] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 753.778015] env[63021]: DEBUG nova.network.neutron [-] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 753.796981] env[63021]: DEBUG nova.network.neutron [-] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 753.920593] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.455s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.921193] env[63021]: DEBUG nova.compute.manager [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 753.924320] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.315s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.925744] env[63021]: INFO nova.compute.claims [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 754.074714] env[63021]: DEBUG nova.network.neutron [req-adae83f7-da48-4003-a03c-08b535a71e54 req-bbe2c37e-1fde-4e2a-8032-7f1e004974f0 service nova] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 754.190555] env[63021]: DEBUG nova.network.neutron [req-adae83f7-da48-4003-a03c-08b535a71e54 req-bbe2c37e-1fde-4e2a-8032-7f1e004974f0 service nova] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.300015] env[63021]: DEBUG nova.network.neutron [-] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.429746] env[63021]: DEBUG nova.compute.utils [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 754.433485] env[63021]: DEBUG nova.compute.manager [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 754.433689] env[63021]: DEBUG nova.network.neutron [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 754.482766] env[63021]: DEBUG nova.policy [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cfc48d1f237f4597ad06f9d7a6bd8a00', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4b4c8cd929ff4068b45ee829b19f9769', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 754.694018] env[63021]: DEBUG oslo_concurrency.lockutils [req-adae83f7-da48-4003-a03c-08b535a71e54 req-bbe2c37e-1fde-4e2a-8032-7f1e004974f0 service nova] Releasing lock "refresh_cache-907540da-a701-477a-9b5c-9942b5d2d987" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.694133] env[63021]: DEBUG nova.compute.manager [req-adae83f7-da48-4003-a03c-08b535a71e54 req-bbe2c37e-1fde-4e2a-8032-7f1e004974f0 service nova] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Received event network-vif-deleted-5cea81bb-fd5c-430e-b8f1-5013d52a6802 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 754.694412] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquired lock "refresh_cache-907540da-a701-477a-9b5c-9942b5d2d987" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.695028] env[63021]: DEBUG nova.network.neutron [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 754.787681] env[63021]: DEBUG nova.network.neutron [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Successfully created port: c4616f5a-0905-4083-83b8-33fb1d8dd324 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 754.802427] env[63021]: INFO nova.compute.manager [-] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Took 1.02 seconds to deallocate network for instance. [ 754.804709] env[63021]: DEBUG nova.compute.claims [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 754.804876] env[63021]: DEBUG oslo_concurrency.lockutils [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.936323] env[63021]: DEBUG nova.compute.manager [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 755.217669] env[63021]: DEBUG nova.network.neutron [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 755.248572] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6cffc17-844e-4ffc-897d-b84e1c843624 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.259283] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24b662a2-5540-4671-b6c2-1efbcb8fd5bf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.292115] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3739cb57-383b-41c4-8b2a-0f123f34ab2a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.301503] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d374e06-12a9-407f-a0b6-b831322e7681 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.317640] env[63021]: DEBUG nova.compute.provider_tree [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 755.337527] env[63021]: DEBUG nova.network.neutron [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.820975] env[63021]: DEBUG nova.scheduler.client.report [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 755.837406] env[63021]: DEBUG nova.compute.manager [req-33e8dca2-3f31-4f48-8f13-0bae226459a1 req-9d24e6f1-920d-42b0-ad93-c8f93e3f8977 service nova] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Received event network-changed-c4616f5a-0905-4083-83b8-33fb1d8dd324 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 755.837406] env[63021]: DEBUG nova.compute.manager [req-33e8dca2-3f31-4f48-8f13-0bae226459a1 req-9d24e6f1-920d-42b0-ad93-c8f93e3f8977 service nova] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Refreshing instance network info cache due to event network-changed-c4616f5a-0905-4083-83b8-33fb1d8dd324. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 755.837406] env[63021]: DEBUG oslo_concurrency.lockutils [req-33e8dca2-3f31-4f48-8f13-0bae226459a1 req-9d24e6f1-920d-42b0-ad93-c8f93e3f8977 service nova] Acquiring lock "refresh_cache-06ef77f1-a196-499c-b639-8427b021a8c0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.837406] env[63021]: DEBUG oslo_concurrency.lockutils [req-33e8dca2-3f31-4f48-8f13-0bae226459a1 req-9d24e6f1-920d-42b0-ad93-c8f93e3f8977 service nova] Acquired lock "refresh_cache-06ef77f1-a196-499c-b639-8427b021a8c0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.837406] env[63021]: DEBUG nova.network.neutron [req-33e8dca2-3f31-4f48-8f13-0bae226459a1 req-9d24e6f1-920d-42b0-ad93-c8f93e3f8977 service nova] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Refreshing network info cache for port c4616f5a-0905-4083-83b8-33fb1d8dd324 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 755.839642] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Releasing lock "refresh_cache-907540da-a701-477a-9b5c-9942b5d2d987" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.840018] env[63021]: DEBUG nova.compute.manager [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 755.840213] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 755.840705] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-25c9186c-8368-4cae-9179-653623a1cda7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.850662] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24dd338f-bd3f-4a1d-8ef6-31f575669eae {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.871668] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 907540da-a701-477a-9b5c-9942b5d2d987 could not be found. [ 755.871952] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 755.872059] env[63021]: INFO nova.compute.manager [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Took 0.03 seconds to destroy the instance on the hypervisor. [ 755.872297] env[63021]: DEBUG oslo.service.loopingcall [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 755.872506] env[63021]: DEBUG nova.compute.manager [-] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 755.872597] env[63021]: DEBUG nova.network.neutron [-] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 755.888428] env[63021]: DEBUG nova.network.neutron [-] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 755.954107] env[63021]: DEBUG nova.compute.manager [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 755.980931] env[63021]: DEBUG nova.virt.hardware [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 755.981193] env[63021]: DEBUG nova.virt.hardware [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 755.981586] env[63021]: DEBUG nova.virt.hardware [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 755.981926] env[63021]: DEBUG nova.virt.hardware [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 755.982162] env[63021]: DEBUG nova.virt.hardware [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 755.982412] env[63021]: DEBUG nova.virt.hardware [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 755.982750] env[63021]: DEBUG nova.virt.hardware [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 755.982979] env[63021]: DEBUG nova.virt.hardware [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 755.983071] env[63021]: DEBUG nova.virt.hardware [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 755.983256] env[63021]: DEBUG nova.virt.hardware [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 755.983442] env[63021]: DEBUG nova.virt.hardware [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 755.984387] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5046dc8e-d28b-4e9f-8289-19c3efa0ee9d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.992477] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6535c128-269f-4ccd-a0c6-126109dd96ca {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.020298] env[63021]: ERROR nova.compute.manager [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c4616f5a-0905-4083-83b8-33fb1d8dd324, please check neutron logs for more information. [ 756.020298] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 756.020298] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 756.020298] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 756.020298] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 756.020298] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 756.020298] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 756.020298] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 756.020298] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 756.020298] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 756.020298] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 756.020298] env[63021]: ERROR nova.compute.manager raise self.value [ 756.020298] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 756.020298] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 756.020298] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 756.020298] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 756.020876] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 756.020876] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 756.020876] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c4616f5a-0905-4083-83b8-33fb1d8dd324, please check neutron logs for more information. [ 756.020876] env[63021]: ERROR nova.compute.manager [ 756.020876] env[63021]: Traceback (most recent call last): [ 756.020876] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 756.020876] env[63021]: listener.cb(fileno) [ 756.020876] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 756.020876] env[63021]: result = function(*args, **kwargs) [ 756.020876] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 756.020876] env[63021]: return func(*args, **kwargs) [ 756.020876] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 756.020876] env[63021]: raise e [ 756.020876] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 756.020876] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 756.020876] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 756.020876] env[63021]: created_port_ids = self._update_ports_for_instance( [ 756.020876] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 756.020876] env[63021]: with excutils.save_and_reraise_exception(): [ 756.020876] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 756.020876] env[63021]: self.force_reraise() [ 756.020876] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 756.020876] env[63021]: raise self.value [ 756.020876] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 756.020876] env[63021]: updated_port = self._update_port( [ 756.020876] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 756.020876] env[63021]: _ensure_no_port_binding_failure(port) [ 756.020876] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 756.020876] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 756.021819] env[63021]: nova.exception.PortBindingFailed: Binding failed for port c4616f5a-0905-4083-83b8-33fb1d8dd324, please check neutron logs for more information. [ 756.021819] env[63021]: Removing descriptor: 17 [ 756.021819] env[63021]: ERROR nova.compute.manager [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c4616f5a-0905-4083-83b8-33fb1d8dd324, please check neutron logs for more information. [ 756.021819] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Traceback (most recent call last): [ 756.021819] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 756.021819] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] yield resources [ 756.021819] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 756.021819] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] self.driver.spawn(context, instance, image_meta, [ 756.021819] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 756.021819] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 756.021819] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 756.021819] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] vm_ref = self.build_virtual_machine(instance, [ 756.022238] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 756.022238] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] vif_infos = vmwarevif.get_vif_info(self._session, [ 756.022238] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 756.022238] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] for vif in network_info: [ 756.022238] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 756.022238] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] return self._sync_wrapper(fn, *args, **kwargs) [ 756.022238] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 756.022238] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] self.wait() [ 756.022238] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 756.022238] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] self[:] = self._gt.wait() [ 756.022238] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 756.022238] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] return self._exit_event.wait() [ 756.022238] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 756.022664] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] result = hub.switch() [ 756.022664] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 756.022664] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] return self.greenlet.switch() [ 756.022664] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 756.022664] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] result = function(*args, **kwargs) [ 756.022664] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 756.022664] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] return func(*args, **kwargs) [ 756.022664] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 756.022664] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] raise e [ 756.022664] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 756.022664] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] nwinfo = self.network_api.allocate_for_instance( [ 756.022664] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 756.022664] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] created_port_ids = self._update_ports_for_instance( [ 756.023150] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 756.023150] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] with excutils.save_and_reraise_exception(): [ 756.023150] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 756.023150] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] self.force_reraise() [ 756.023150] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 756.023150] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] raise self.value [ 756.023150] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 756.023150] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] updated_port = self._update_port( [ 756.023150] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 756.023150] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] _ensure_no_port_binding_failure(port) [ 756.023150] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 756.023150] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] raise exception.PortBindingFailed(port_id=port['id']) [ 756.023550] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] nova.exception.PortBindingFailed: Binding failed for port c4616f5a-0905-4083-83b8-33fb1d8dd324, please check neutron logs for more information. [ 756.023550] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] [ 756.023550] env[63021]: INFO nova.compute.manager [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Terminating instance [ 756.024505] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Acquiring lock "refresh_cache-06ef77f1-a196-499c-b639-8427b021a8c0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.326085] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.401s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.326447] env[63021]: DEBUG nova.compute.manager [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 756.328951] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.827s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.330283] env[63021]: INFO nova.compute.claims [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 756.352646] env[63021]: DEBUG nova.network.neutron [req-33e8dca2-3f31-4f48-8f13-0bae226459a1 req-9d24e6f1-920d-42b0-ad93-c8f93e3f8977 service nova] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 756.391117] env[63021]: DEBUG nova.network.neutron [-] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.440917] env[63021]: DEBUG nova.network.neutron [req-33e8dca2-3f31-4f48-8f13-0bae226459a1 req-9d24e6f1-920d-42b0-ad93-c8f93e3f8977 service nova] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.834593] env[63021]: DEBUG nova.compute.utils [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 756.837907] env[63021]: DEBUG nova.compute.manager [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 756.838092] env[63021]: DEBUG nova.network.neutron [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 756.893320] env[63021]: DEBUG nova.policy [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bc932171c4cc4c11908d7d758c11fa69', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f7c6a79591b342d69540fa4b07c7d7ce', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 756.895110] env[63021]: INFO nova.compute.manager [-] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Took 1.02 seconds to deallocate network for instance. [ 756.897613] env[63021]: DEBUG nova.compute.claims [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 756.897849] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.944190] env[63021]: DEBUG oslo_concurrency.lockutils [req-33e8dca2-3f31-4f48-8f13-0bae226459a1 req-9d24e6f1-920d-42b0-ad93-c8f93e3f8977 service nova] Releasing lock "refresh_cache-06ef77f1-a196-499c-b639-8427b021a8c0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.944706] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Acquired lock "refresh_cache-06ef77f1-a196-499c-b639-8427b021a8c0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.944962] env[63021]: DEBUG nova.network.neutron [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 757.338461] env[63021]: DEBUG nova.compute.manager [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 757.348615] env[63021]: DEBUG nova.network.neutron [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Successfully created port: 08a404ab-ec2a-4859-b302-f163e69680c2 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 757.464970] env[63021]: DEBUG nova.network.neutron [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 757.544725] env[63021]: DEBUG nova.network.neutron [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.634190] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a85a49fe-fee2-4a17-b200-f094d057520c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.642778] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85ea9055-0d3f-4928-81d0-e93f86d6944e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.671304] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-877cbb4a-c88f-4b6f-9c80-57bec5ff0a5a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.679642] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63b983d7-9030-4a64-92a6-ab1a24e4eb74 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.693547] env[63021]: DEBUG nova.compute.provider_tree [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 757.929587] env[63021]: DEBUG nova.compute.manager [req-b010c6f7-fdd6-4cea-9a9c-e7ffc19192fd req-ba0f1aa1-e94a-421a-a22d-14e59ad81737 service nova] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Received event network-vif-deleted-c4616f5a-0905-4083-83b8-33fb1d8dd324 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 758.049674] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Releasing lock "refresh_cache-06ef77f1-a196-499c-b639-8427b021a8c0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.050175] env[63021]: DEBUG nova.compute.manager [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 758.050374] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 758.050752] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-386827d1-7473-4bbd-a502-8ad4da86752b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.060290] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae28ec2c-7e2e-4efb-b56b-c64345cc9e9f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.084278] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 06ef77f1-a196-499c-b639-8427b021a8c0 could not be found. [ 758.084501] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 758.084676] env[63021]: INFO nova.compute.manager [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Took 0.03 seconds to destroy the instance on the hypervisor. [ 758.084919] env[63021]: DEBUG oslo.service.loopingcall [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 758.085151] env[63021]: DEBUG nova.compute.manager [-] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 758.085249] env[63021]: DEBUG nova.network.neutron [-] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 758.100171] env[63021]: DEBUG nova.network.neutron [-] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 758.195671] env[63021]: DEBUG nova.scheduler.client.report [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 758.351395] env[63021]: DEBUG nova.compute.manager [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 758.383894] env[63021]: DEBUG nova.virt.hardware [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 758.384158] env[63021]: DEBUG nova.virt.hardware [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 758.384852] env[63021]: DEBUG nova.virt.hardware [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 758.385062] env[63021]: DEBUG nova.virt.hardware [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 758.385212] env[63021]: DEBUG nova.virt.hardware [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 758.385358] env[63021]: DEBUG nova.virt.hardware [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 758.385562] env[63021]: DEBUG nova.virt.hardware [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 758.385724] env[63021]: DEBUG nova.virt.hardware [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 758.385886] env[63021]: DEBUG nova.virt.hardware [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 758.386055] env[63021]: DEBUG nova.virt.hardware [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 758.386230] env[63021]: DEBUG nova.virt.hardware [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 758.387859] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee5a1dfe-f62e-4237-b955-3a89d8d39ffa {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.396493] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45b85754-4631-4b8b-9100-bec92fe97ebe {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.420525] env[63021]: ERROR nova.compute.manager [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 08a404ab-ec2a-4859-b302-f163e69680c2, please check neutron logs for more information. [ 758.420525] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 758.420525] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 758.420525] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 758.420525] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 758.420525] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 758.420525] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 758.420525] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 758.420525] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.420525] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 758.420525] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.420525] env[63021]: ERROR nova.compute.manager raise self.value [ 758.420525] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 758.420525] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 758.420525] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.420525] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 758.421334] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.421334] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 758.421334] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 08a404ab-ec2a-4859-b302-f163e69680c2, please check neutron logs for more information. [ 758.421334] env[63021]: ERROR nova.compute.manager [ 758.421334] env[63021]: Traceback (most recent call last): [ 758.421334] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 758.421334] env[63021]: listener.cb(fileno) [ 758.421334] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 758.421334] env[63021]: result = function(*args, **kwargs) [ 758.421334] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 758.421334] env[63021]: return func(*args, **kwargs) [ 758.421334] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 758.421334] env[63021]: raise e [ 758.421334] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 758.421334] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 758.421334] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 758.421334] env[63021]: created_port_ids = self._update_ports_for_instance( [ 758.421334] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 758.421334] env[63021]: with excutils.save_and_reraise_exception(): [ 758.421334] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.421334] env[63021]: self.force_reraise() [ 758.421334] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.421334] env[63021]: raise self.value [ 758.421334] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 758.421334] env[63021]: updated_port = self._update_port( [ 758.421334] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.421334] env[63021]: _ensure_no_port_binding_failure(port) [ 758.421334] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.421334] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 758.423542] env[63021]: nova.exception.PortBindingFailed: Binding failed for port 08a404ab-ec2a-4859-b302-f163e69680c2, please check neutron logs for more information. [ 758.423542] env[63021]: Removing descriptor: 17 [ 758.423542] env[63021]: ERROR nova.compute.manager [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 08a404ab-ec2a-4859-b302-f163e69680c2, please check neutron logs for more information. [ 758.423542] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Traceback (most recent call last): [ 758.423542] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 758.423542] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] yield resources [ 758.423542] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 758.423542] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] self.driver.spawn(context, instance, image_meta, [ 758.423542] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 758.423542] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 758.423542] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 758.423542] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] vm_ref = self.build_virtual_machine(instance, [ 758.424041] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 758.424041] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] vif_infos = vmwarevif.get_vif_info(self._session, [ 758.424041] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 758.424041] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] for vif in network_info: [ 758.424041] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 758.424041] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] return self._sync_wrapper(fn, *args, **kwargs) [ 758.424041] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 758.424041] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] self.wait() [ 758.424041] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 758.424041] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] self[:] = self._gt.wait() [ 758.424041] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 758.424041] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] return self._exit_event.wait() [ 758.424041] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 758.424504] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] result = hub.switch() [ 758.424504] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 758.424504] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] return self.greenlet.switch() [ 758.424504] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 758.424504] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] result = function(*args, **kwargs) [ 758.424504] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 758.424504] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] return func(*args, **kwargs) [ 758.424504] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 758.424504] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] raise e [ 758.424504] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 758.424504] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] nwinfo = self.network_api.allocate_for_instance( [ 758.424504] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 758.424504] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] created_port_ids = self._update_ports_for_instance( [ 758.424970] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 758.424970] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] with excutils.save_and_reraise_exception(): [ 758.424970] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.424970] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] self.force_reraise() [ 758.424970] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.424970] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] raise self.value [ 758.424970] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 758.424970] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] updated_port = self._update_port( [ 758.424970] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.424970] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] _ensure_no_port_binding_failure(port) [ 758.424970] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.424970] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] raise exception.PortBindingFailed(port_id=port['id']) [ 758.425398] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] nova.exception.PortBindingFailed: Binding failed for port 08a404ab-ec2a-4859-b302-f163e69680c2, please check neutron logs for more information. [ 758.425398] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] [ 758.425398] env[63021]: INFO nova.compute.manager [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Terminating instance [ 758.425398] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Acquiring lock "refresh_cache-8ea2a693-0cd8-4f91-8bd8-b40265b9372e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.425398] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Acquired lock "refresh_cache-8ea2a693-0cd8-4f91-8bd8-b40265b9372e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.425398] env[63021]: DEBUG nova.network.neutron [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 758.603045] env[63021]: DEBUG nova.network.neutron [-] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.700862] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.372s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.701445] env[63021]: DEBUG nova.compute.manager [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 758.704191] env[63021]: DEBUG oslo_concurrency.lockutils [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.140s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 758.944093] env[63021]: DEBUG nova.network.neutron [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.035987] env[63021]: DEBUG nova.network.neutron [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.106050] env[63021]: INFO nova.compute.manager [-] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Took 1.02 seconds to deallocate network for instance. [ 759.108488] env[63021]: DEBUG nova.compute.claims [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 759.108678] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.209071] env[63021]: DEBUG nova.compute.utils [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 759.213667] env[63021]: DEBUG nova.compute.manager [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 759.213918] env[63021]: DEBUG nova.network.neutron [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 759.255565] env[63021]: DEBUG nova.policy [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2aedf7c17a8e4a4caf7285f183f4f731', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '87484a3f9cf94549a4290f436544506f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 759.529320] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ab987ab-84cb-442c-a0c5-d840a4fd2a8a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.536819] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79d0212f-902f-4b57-94cf-5e18442f0458 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.541705] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Releasing lock "refresh_cache-8ea2a693-0cd8-4f91-8bd8-b40265b9372e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.542156] env[63021]: DEBUG nova.compute.manager [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 759.542345] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 759.542587] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-67c3698e-69a1-48dd-b79a-dc632c5f09d5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.571592] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fff25316-ece2-4362-8840-ba39cd8d50e8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.576507] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20d5fdd0-26b8-4429-a6cc-6fc704b63e6a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.592055] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-752d4381-d045-4276-a9b0-0af0a03596de {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.599420] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8ea2a693-0cd8-4f91-8bd8-b40265b9372e could not be found. [ 759.599420] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 759.599420] env[63021]: INFO nova.compute.manager [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Took 0.06 seconds to destroy the instance on the hypervisor. [ 759.599659] env[63021]: DEBUG oslo.service.loopingcall [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 759.600121] env[63021]: DEBUG nova.compute.manager [-] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 759.600296] env[63021]: DEBUG nova.network.neutron [-] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 759.608944] env[63021]: DEBUG nova.compute.provider_tree [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 759.610622] env[63021]: DEBUG nova.network.neutron [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Successfully created port: 6cab39b0-f517-48fa-916b-2a14b0be3861 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 759.624520] env[63021]: DEBUG nova.network.neutron [-] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.717318] env[63021]: DEBUG nova.compute.manager [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 759.885436] env[63021]: DEBUG nova.network.neutron [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Successfully created port: ff06999c-1d72-463d-a22e-db95a77de588 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 759.981586] env[63021]: DEBUG nova.compute.manager [req-34bb64b4-5ee4-4a9b-862f-89bc23046bf7 req-4b1af26e-16d4-4ca1-a77f-ce745f387d94 service nova] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Received event network-changed-08a404ab-ec2a-4859-b302-f163e69680c2 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 759.981704] env[63021]: DEBUG nova.compute.manager [req-34bb64b4-5ee4-4a9b-862f-89bc23046bf7 req-4b1af26e-16d4-4ca1-a77f-ce745f387d94 service nova] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Refreshing instance network info cache due to event network-changed-08a404ab-ec2a-4859-b302-f163e69680c2. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 759.981913] env[63021]: DEBUG oslo_concurrency.lockutils [req-34bb64b4-5ee4-4a9b-862f-89bc23046bf7 req-4b1af26e-16d4-4ca1-a77f-ce745f387d94 service nova] Acquiring lock "refresh_cache-8ea2a693-0cd8-4f91-8bd8-b40265b9372e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.982160] env[63021]: DEBUG oslo_concurrency.lockutils [req-34bb64b4-5ee4-4a9b-862f-89bc23046bf7 req-4b1af26e-16d4-4ca1-a77f-ce745f387d94 service nova] Acquired lock "refresh_cache-8ea2a693-0cd8-4f91-8bd8-b40265b9372e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.982245] env[63021]: DEBUG nova.network.neutron [req-34bb64b4-5ee4-4a9b-862f-89bc23046bf7 req-4b1af26e-16d4-4ca1-a77f-ce745f387d94 service nova] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Refreshing network info cache for port 08a404ab-ec2a-4859-b302-f163e69680c2 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 760.113840] env[63021]: DEBUG nova.scheduler.client.report [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 760.126568] env[63021]: DEBUG nova.network.neutron [-] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.500749] env[63021]: DEBUG nova.network.neutron [req-34bb64b4-5ee4-4a9b-862f-89bc23046bf7 req-4b1af26e-16d4-4ca1-a77f-ce745f387d94 service nova] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 760.574147] env[63021]: DEBUG nova.network.neutron [req-34bb64b4-5ee4-4a9b-862f-89bc23046bf7 req-4b1af26e-16d4-4ca1-a77f-ce745f387d94 service nova] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.618877] env[63021]: DEBUG oslo_concurrency.lockutils [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.915s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.619486] env[63021]: ERROR nova.compute.manager [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port dda3fa9a-a2cf-4db2-a54f-f5cfe182e810, please check neutron logs for more information. [ 760.619486] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] Traceback (most recent call last): [ 760.619486] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 760.619486] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] self.driver.spawn(context, instance, image_meta, [ 760.619486] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 760.619486] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 760.619486] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 760.619486] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] vm_ref = self.build_virtual_machine(instance, [ 760.619486] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 760.619486] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] vif_infos = vmwarevif.get_vif_info(self._session, [ 760.619486] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 760.619884] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] for vif in network_info: [ 760.619884] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 760.619884] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] return self._sync_wrapper(fn, *args, **kwargs) [ 760.619884] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 760.619884] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] self.wait() [ 760.619884] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 760.619884] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] self[:] = self._gt.wait() [ 760.619884] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 760.619884] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] return self._exit_event.wait() [ 760.619884] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 760.619884] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] result = hub.switch() [ 760.619884] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 760.619884] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] return self.greenlet.switch() [ 760.620393] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 760.620393] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] result = function(*args, **kwargs) [ 760.620393] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 760.620393] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] return func(*args, **kwargs) [ 760.620393] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 760.620393] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] raise e [ 760.620393] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 760.620393] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] nwinfo = self.network_api.allocate_for_instance( [ 760.620393] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 760.620393] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] created_port_ids = self._update_ports_for_instance( [ 760.620393] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 760.620393] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] with excutils.save_and_reraise_exception(): [ 760.620393] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 760.620840] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] self.force_reraise() [ 760.620840] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 760.620840] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] raise self.value [ 760.620840] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 760.620840] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] updated_port = self._update_port( [ 760.620840] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 760.620840] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] _ensure_no_port_binding_failure(port) [ 760.620840] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 760.620840] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] raise exception.PortBindingFailed(port_id=port['id']) [ 760.620840] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] nova.exception.PortBindingFailed: Binding failed for port dda3fa9a-a2cf-4db2-a54f-f5cfe182e810, please check neutron logs for more information. [ 760.620840] env[63021]: ERROR nova.compute.manager [instance: 274b001a-443f-463c-a657-652f64ba1de8] [ 760.621297] env[63021]: DEBUG nova.compute.utils [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Binding failed for port dda3fa9a-a2cf-4db2-a54f-f5cfe182e810, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 760.621542] env[63021]: DEBUG oslo_concurrency.lockutils [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.668s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.623401] env[63021]: INFO nova.compute.claims [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] [instance: f77da305-9cc9-415d-a808-c2136212c456] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 760.627222] env[63021]: DEBUG nova.compute.manager [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Build of instance 274b001a-443f-463c-a657-652f64ba1de8 was re-scheduled: Binding failed for port dda3fa9a-a2cf-4db2-a54f-f5cfe182e810, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 760.627479] env[63021]: DEBUG nova.compute.manager [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 760.627716] env[63021]: DEBUG oslo_concurrency.lockutils [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Acquiring lock "refresh_cache-274b001a-443f-463c-a657-652f64ba1de8" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.627859] env[63021]: DEBUG oslo_concurrency.lockutils [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Acquired lock "refresh_cache-274b001a-443f-463c-a657-652f64ba1de8" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.628021] env[63021]: DEBUG nova.network.neutron [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 760.629414] env[63021]: INFO nova.compute.manager [-] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Took 1.03 seconds to deallocate network for instance. [ 760.634279] env[63021]: DEBUG nova.compute.claims [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 760.634445] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.727182] env[63021]: DEBUG nova.compute.manager [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 760.751548] env[63021]: DEBUG nova.virt.hardware [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 760.751785] env[63021]: DEBUG nova.virt.hardware [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 760.751937] env[63021]: DEBUG nova.virt.hardware [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 760.752125] env[63021]: DEBUG nova.virt.hardware [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 760.752268] env[63021]: DEBUG nova.virt.hardware [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 760.752409] env[63021]: DEBUG nova.virt.hardware [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 760.752612] env[63021]: DEBUG nova.virt.hardware [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 760.752763] env[63021]: DEBUG nova.virt.hardware [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 760.752990] env[63021]: DEBUG nova.virt.hardware [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 760.753271] env[63021]: DEBUG nova.virt.hardware [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 760.753455] env[63021]: DEBUG nova.virt.hardware [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 760.754385] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bb99eb0-7a70-4f4b-9a8b-cc0434f05f7a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.763066] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26c9ee33-e8e9-4a34-b951-4966230445fd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.076130] env[63021]: DEBUG oslo_concurrency.lockutils [req-34bb64b4-5ee4-4a9b-862f-89bc23046bf7 req-4b1af26e-16d4-4ca1-a77f-ce745f387d94 service nova] Releasing lock "refresh_cache-8ea2a693-0cd8-4f91-8bd8-b40265b9372e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.076444] env[63021]: DEBUG nova.compute.manager [req-34bb64b4-5ee4-4a9b-862f-89bc23046bf7 req-4b1af26e-16d4-4ca1-a77f-ce745f387d94 service nova] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Received event network-vif-deleted-08a404ab-ec2a-4859-b302-f163e69680c2 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 761.145607] env[63021]: ERROR nova.compute.manager [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6cab39b0-f517-48fa-916b-2a14b0be3861, please check neutron logs for more information. [ 761.145607] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 761.145607] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 761.145607] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 761.145607] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 761.145607] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 761.145607] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 761.145607] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 761.145607] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.145607] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 761.145607] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.145607] env[63021]: ERROR nova.compute.manager raise self.value [ 761.145607] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 761.145607] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 761.145607] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.145607] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 761.146129] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.146129] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 761.146129] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6cab39b0-f517-48fa-916b-2a14b0be3861, please check neutron logs for more information. [ 761.146129] env[63021]: ERROR nova.compute.manager [ 761.146129] env[63021]: Traceback (most recent call last): [ 761.146129] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 761.146129] env[63021]: listener.cb(fileno) [ 761.146129] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 761.146129] env[63021]: result = function(*args, **kwargs) [ 761.146129] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 761.146129] env[63021]: return func(*args, **kwargs) [ 761.146129] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 761.146129] env[63021]: raise e [ 761.146129] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 761.146129] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 761.146129] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 761.146129] env[63021]: created_port_ids = self._update_ports_for_instance( [ 761.146129] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 761.146129] env[63021]: with excutils.save_and_reraise_exception(): [ 761.146129] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.146129] env[63021]: self.force_reraise() [ 761.146129] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.146129] env[63021]: raise self.value [ 761.146129] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 761.146129] env[63021]: updated_port = self._update_port( [ 761.146129] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.146129] env[63021]: _ensure_no_port_binding_failure(port) [ 761.146129] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.146129] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 761.147029] env[63021]: nova.exception.PortBindingFailed: Binding failed for port 6cab39b0-f517-48fa-916b-2a14b0be3861, please check neutron logs for more information. [ 761.147029] env[63021]: Removing descriptor: 17 [ 761.147146] env[63021]: ERROR nova.compute.manager [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6cab39b0-f517-48fa-916b-2a14b0be3861, please check neutron logs for more information. [ 761.147146] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Traceback (most recent call last): [ 761.147146] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 761.147146] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] yield resources [ 761.147146] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 761.147146] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] self.driver.spawn(context, instance, image_meta, [ 761.147146] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 761.147146] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] self._vmops.spawn(context, instance, image_meta, injected_files, [ 761.147146] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 761.147146] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] vm_ref = self.build_virtual_machine(instance, [ 761.147146] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 761.147485] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] vif_infos = vmwarevif.get_vif_info(self._session, [ 761.147485] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 761.147485] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] for vif in network_info: [ 761.147485] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 761.147485] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] return self._sync_wrapper(fn, *args, **kwargs) [ 761.147485] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 761.147485] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] self.wait() [ 761.147485] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 761.147485] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] self[:] = self._gt.wait() [ 761.147485] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 761.147485] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] return self._exit_event.wait() [ 761.147485] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 761.147485] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] result = hub.switch() [ 761.147891] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 761.147891] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] return self.greenlet.switch() [ 761.147891] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 761.147891] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] result = function(*args, **kwargs) [ 761.147891] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 761.147891] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] return func(*args, **kwargs) [ 761.147891] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 761.147891] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] raise e [ 761.147891] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 761.147891] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] nwinfo = self.network_api.allocate_for_instance( [ 761.147891] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 761.147891] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] created_port_ids = self._update_ports_for_instance( [ 761.147891] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 761.148280] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] with excutils.save_and_reraise_exception(): [ 761.148280] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.148280] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] self.force_reraise() [ 761.148280] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.148280] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] raise self.value [ 761.148280] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 761.148280] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] updated_port = self._update_port( [ 761.148280] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.148280] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] _ensure_no_port_binding_failure(port) [ 761.148280] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.148280] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] raise exception.PortBindingFailed(port_id=port['id']) [ 761.148280] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] nova.exception.PortBindingFailed: Binding failed for port 6cab39b0-f517-48fa-916b-2a14b0be3861, please check neutron logs for more information. [ 761.148280] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] [ 761.148881] env[63021]: INFO nova.compute.manager [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Terminating instance [ 761.150023] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Acquiring lock "refresh_cache-d02adae0-724e-4dbe-a6c5-d476805d3d57" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.150841] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Acquired lock "refresh_cache-d02adae0-724e-4dbe-a6c5-d476805d3d57" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.151627] env[63021]: DEBUG nova.network.neutron [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 761.154564] env[63021]: DEBUG nova.network.neutron [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.238237] env[63021]: DEBUG nova.network.neutron [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.692286] env[63021]: DEBUG nova.network.neutron [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.741202] env[63021]: DEBUG oslo_concurrency.lockutils [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Releasing lock "refresh_cache-274b001a-443f-463c-a657-652f64ba1de8" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.741202] env[63021]: DEBUG nova.compute.manager [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 761.741445] env[63021]: DEBUG nova.compute.manager [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 761.741569] env[63021]: DEBUG nova.network.neutron [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 274b001a-443f-463c-a657-652f64ba1de8] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 761.759853] env[63021]: DEBUG nova.network.neutron [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.776034] env[63021]: DEBUG nova.network.neutron [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.907409] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bb696e4-abf5-455d-b787-f0a31660114b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.915489] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fef07fb-a02c-4b67-a58e-5952b7ece0ac {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.947045] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5239a3bb-fd09-43f9-8e05-5e4dad0b9b1b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.954428] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b0fe702-a7b1-4b7f-802a-b499964940f2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.967309] env[63021]: DEBUG nova.compute.provider_tree [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 762.045316] env[63021]: DEBUG nova.compute.manager [req-323652b3-d24f-44ad-a4ba-bae2ebd787db req-c36651f0-c36b-4e49-90f8-339f50acdcbb service nova] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Received event network-changed-6cab39b0-f517-48fa-916b-2a14b0be3861 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 762.045451] env[63021]: DEBUG nova.compute.manager [req-323652b3-d24f-44ad-a4ba-bae2ebd787db req-c36651f0-c36b-4e49-90f8-339f50acdcbb service nova] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Refreshing instance network info cache due to event network-changed-6cab39b0-f517-48fa-916b-2a14b0be3861. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 762.045640] env[63021]: DEBUG oslo_concurrency.lockutils [req-323652b3-d24f-44ad-a4ba-bae2ebd787db req-c36651f0-c36b-4e49-90f8-339f50acdcbb service nova] Acquiring lock "refresh_cache-d02adae0-724e-4dbe-a6c5-d476805d3d57" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.261960] env[63021]: DEBUG nova.network.neutron [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.278064] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Releasing lock "refresh_cache-d02adae0-724e-4dbe-a6c5-d476805d3d57" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.278517] env[63021]: DEBUG nova.compute.manager [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 762.278667] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 762.278958] env[63021]: DEBUG oslo_concurrency.lockutils [req-323652b3-d24f-44ad-a4ba-bae2ebd787db req-c36651f0-c36b-4e49-90f8-339f50acdcbb service nova] Acquired lock "refresh_cache-d02adae0-724e-4dbe-a6c5-d476805d3d57" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.279145] env[63021]: DEBUG nova.network.neutron [req-323652b3-d24f-44ad-a4ba-bae2ebd787db req-c36651f0-c36b-4e49-90f8-339f50acdcbb service nova] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Refreshing network info cache for port 6cab39b0-f517-48fa-916b-2a14b0be3861 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 762.280223] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b8fc500a-3a88-4a95-8a0d-f85044873b70 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.291075] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeca0a30-0d83-4426-95c2-929b047a767c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.314183] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d02adae0-724e-4dbe-a6c5-d476805d3d57 could not be found. [ 762.314520] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 762.314808] env[63021]: INFO nova.compute.manager [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Took 0.04 seconds to destroy the instance on the hypervisor. [ 762.315845] env[63021]: DEBUG oslo.service.loopingcall [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 762.315845] env[63021]: DEBUG nova.compute.manager [-] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 762.315845] env[63021]: DEBUG nova.network.neutron [-] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 762.374507] env[63021]: DEBUG nova.network.neutron [-] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 762.470954] env[63021]: DEBUG nova.scheduler.client.report [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 762.764692] env[63021]: INFO nova.compute.manager [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 274b001a-443f-463c-a657-652f64ba1de8] Took 1.02 seconds to deallocate network for instance. [ 762.799884] env[63021]: DEBUG nova.network.neutron [req-323652b3-d24f-44ad-a4ba-bae2ebd787db req-c36651f0-c36b-4e49-90f8-339f50acdcbb service nova] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 762.875052] env[63021]: DEBUG nova.network.neutron [req-323652b3-d24f-44ad-a4ba-bae2ebd787db req-c36651f0-c36b-4e49-90f8-339f50acdcbb service nova] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.976594] env[63021]: DEBUG oslo_concurrency.lockutils [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.355s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.977088] env[63021]: DEBUG nova.compute.manager [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] [instance: f77da305-9cc9-415d-a808-c2136212c456] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 762.979561] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 16.945s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.979732] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.979931] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63021) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 762.980212] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.054s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.981667] env[63021]: INFO nova.compute.claims [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 762.984973] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-132b219d-d0f5-436f-aaf2-b683b35d78b6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.993460] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d31215a4-4133-4d5a-bc17-9178a83d2e68 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.007273] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19aa07b3-0954-4d7d-8acd-96a655d9037f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.015029] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dd8486d-6060-43d2-8d88-4fcb5c2a2dde {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.044853] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181497MB free_disk=139GB free_vcpus=48 pci_devices=None {{(pid=63021) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 763.045018] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.051084] env[63021]: DEBUG nova.network.neutron [-] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.377781] env[63021]: DEBUG oslo_concurrency.lockutils [req-323652b3-d24f-44ad-a4ba-bae2ebd787db req-c36651f0-c36b-4e49-90f8-339f50acdcbb service nova] Releasing lock "refresh_cache-d02adae0-724e-4dbe-a6c5-d476805d3d57" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.378310] env[63021]: DEBUG nova.compute.manager [req-323652b3-d24f-44ad-a4ba-bae2ebd787db req-c36651f0-c36b-4e49-90f8-339f50acdcbb service nova] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Received event network-vif-deleted-6cab39b0-f517-48fa-916b-2a14b0be3861 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 763.486290] env[63021]: DEBUG nova.compute.utils [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 763.487703] env[63021]: DEBUG nova.compute.manager [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] [instance: f77da305-9cc9-415d-a808-c2136212c456] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 763.487882] env[63021]: DEBUG nova.network.neutron [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] [instance: f77da305-9cc9-415d-a808-c2136212c456] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 763.538388] env[63021]: DEBUG nova.policy [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'be3d91c8bbf94657a202b2896ba8a98e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '454c67ad6bd545e1b4632e8e4e909955', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 763.554464] env[63021]: INFO nova.compute.manager [-] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Took 1.24 seconds to deallocate network for instance. [ 763.556755] env[63021]: DEBUG nova.compute.claims [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 763.557207] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.785759] env[63021]: DEBUG nova.network.neutron [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] [instance: f77da305-9cc9-415d-a808-c2136212c456] Successfully created port: 427f6f68-d9be-4108-b3aa-612f9786561c {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 763.800646] env[63021]: INFO nova.scheduler.client.report [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Deleted allocations for instance 274b001a-443f-463c-a657-652f64ba1de8 [ 763.992498] env[63021]: DEBUG nova.compute.manager [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] [instance: f77da305-9cc9-415d-a808-c2136212c456] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 764.301017] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a041eda1-2276-445f-a46d-c500fa3df3f8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.310437] env[63021]: DEBUG oslo_concurrency.lockutils [None req-521846bb-fd1b-40a4-9135-7c9e2ea37c84 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Lock "274b001a-443f-463c-a657-652f64ba1de8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 130.874s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.312603] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-780af2f0-ff3d-475a-897e-21fd4b7679e8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.346332] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38474fe0-e618-4862-8a9d-e514aecf0237 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.356029] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f532bd74-17a0-4025-bdfb-c37022fb5b55 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.371411] env[63021]: DEBUG nova.compute.provider_tree [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 764.465223] env[63021]: DEBUG nova.compute.manager [req-78cc699e-0fda-4ae2-93e4-fea3152a8364 req-116ff21a-395d-4472-b095-dd071e8f656d service nova] [instance: f77da305-9cc9-415d-a808-c2136212c456] Received event network-changed-427f6f68-d9be-4108-b3aa-612f9786561c {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 764.465223] env[63021]: DEBUG nova.compute.manager [req-78cc699e-0fda-4ae2-93e4-fea3152a8364 req-116ff21a-395d-4472-b095-dd071e8f656d service nova] [instance: f77da305-9cc9-415d-a808-c2136212c456] Refreshing instance network info cache due to event network-changed-427f6f68-d9be-4108-b3aa-612f9786561c. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 764.465423] env[63021]: DEBUG oslo_concurrency.lockutils [req-78cc699e-0fda-4ae2-93e4-fea3152a8364 req-116ff21a-395d-4472-b095-dd071e8f656d service nova] Acquiring lock "refresh_cache-f77da305-9cc9-415d-a808-c2136212c456" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.465556] env[63021]: DEBUG oslo_concurrency.lockutils [req-78cc699e-0fda-4ae2-93e4-fea3152a8364 req-116ff21a-395d-4472-b095-dd071e8f656d service nova] Acquired lock "refresh_cache-f77da305-9cc9-415d-a808-c2136212c456" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.465704] env[63021]: DEBUG nova.network.neutron [req-78cc699e-0fda-4ae2-93e4-fea3152a8364 req-116ff21a-395d-4472-b095-dd071e8f656d service nova] [instance: f77da305-9cc9-415d-a808-c2136212c456] Refreshing network info cache for port 427f6f68-d9be-4108-b3aa-612f9786561c {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 764.606706] env[63021]: ERROR nova.compute.manager [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 427f6f68-d9be-4108-b3aa-612f9786561c, please check neutron logs for more information. [ 764.606706] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 764.606706] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 764.606706] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 764.606706] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 764.606706] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 764.606706] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 764.606706] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 764.606706] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 764.606706] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 764.606706] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 764.606706] env[63021]: ERROR nova.compute.manager raise self.value [ 764.606706] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 764.606706] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 764.606706] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 764.606706] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 764.607296] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 764.607296] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 764.607296] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 427f6f68-d9be-4108-b3aa-612f9786561c, please check neutron logs for more information. [ 764.607296] env[63021]: ERROR nova.compute.manager [ 764.607296] env[63021]: Traceback (most recent call last): [ 764.607296] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 764.607296] env[63021]: listener.cb(fileno) [ 764.607296] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 764.607296] env[63021]: result = function(*args, **kwargs) [ 764.607296] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 764.607296] env[63021]: return func(*args, **kwargs) [ 764.607296] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 764.607296] env[63021]: raise e [ 764.607296] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 764.607296] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 764.607296] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 764.607296] env[63021]: created_port_ids = self._update_ports_for_instance( [ 764.607296] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 764.607296] env[63021]: with excutils.save_and_reraise_exception(): [ 764.607296] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 764.607296] env[63021]: self.force_reraise() [ 764.607296] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 764.607296] env[63021]: raise self.value [ 764.607296] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 764.607296] env[63021]: updated_port = self._update_port( [ 764.607296] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 764.607296] env[63021]: _ensure_no_port_binding_failure(port) [ 764.607296] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 764.607296] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 764.608306] env[63021]: nova.exception.PortBindingFailed: Binding failed for port 427f6f68-d9be-4108-b3aa-612f9786561c, please check neutron logs for more information. [ 764.608306] env[63021]: Removing descriptor: 18 [ 764.817794] env[63021]: DEBUG nova.compute.manager [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 764.875243] env[63021]: DEBUG nova.scheduler.client.report [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 764.983548] env[63021]: DEBUG nova.network.neutron [req-78cc699e-0fda-4ae2-93e4-fea3152a8364 req-116ff21a-395d-4472-b095-dd071e8f656d service nova] [instance: f77da305-9cc9-415d-a808-c2136212c456] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 765.005042] env[63021]: DEBUG nova.compute.manager [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] [instance: f77da305-9cc9-415d-a808-c2136212c456] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 765.030338] env[63021]: DEBUG nova.virt.hardware [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 765.030624] env[63021]: DEBUG nova.virt.hardware [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 765.030780] env[63021]: DEBUG nova.virt.hardware [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 765.030958] env[63021]: DEBUG nova.virt.hardware [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 765.031150] env[63021]: DEBUG nova.virt.hardware [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 765.031300] env[63021]: DEBUG nova.virt.hardware [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 765.031501] env[63021]: DEBUG nova.virt.hardware [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 765.031656] env[63021]: DEBUG nova.virt.hardware [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 765.031855] env[63021]: DEBUG nova.virt.hardware [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 765.032062] env[63021]: DEBUG nova.virt.hardware [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 765.032205] env[63021]: DEBUG nova.virt.hardware [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 765.033088] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f458327-d691-4420-a273-e1703dcdc3bb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.041056] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baab2a1d-6ae2-4e95-9b5c-e52eeede3810 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.054747] env[63021]: ERROR nova.compute.manager [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] [instance: f77da305-9cc9-415d-a808-c2136212c456] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 427f6f68-d9be-4108-b3aa-612f9786561c, please check neutron logs for more information. [ 765.054747] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] Traceback (most recent call last): [ 765.054747] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 765.054747] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] yield resources [ 765.054747] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 765.054747] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] self.driver.spawn(context, instance, image_meta, [ 765.054747] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 765.054747] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] self._vmops.spawn(context, instance, image_meta, injected_files, [ 765.054747] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 765.054747] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] vm_ref = self.build_virtual_machine(instance, [ 765.054747] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 765.055246] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] vif_infos = vmwarevif.get_vif_info(self._session, [ 765.055246] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 765.055246] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] for vif in network_info: [ 765.055246] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 765.055246] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] return self._sync_wrapper(fn, *args, **kwargs) [ 765.055246] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 765.055246] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] self.wait() [ 765.055246] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 765.055246] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] self[:] = self._gt.wait() [ 765.055246] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 765.055246] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] return self._exit_event.wait() [ 765.055246] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 765.055246] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] current.throw(*self._exc) [ 765.055668] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 765.055668] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] result = function(*args, **kwargs) [ 765.055668] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 765.055668] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] return func(*args, **kwargs) [ 765.055668] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 765.055668] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] raise e [ 765.055668] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 765.055668] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] nwinfo = self.network_api.allocate_for_instance( [ 765.055668] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 765.055668] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] created_port_ids = self._update_ports_for_instance( [ 765.055668] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 765.055668] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] with excutils.save_and_reraise_exception(): [ 765.055668] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 765.056180] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] self.force_reraise() [ 765.056180] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 765.056180] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] raise self.value [ 765.056180] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 765.056180] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] updated_port = self._update_port( [ 765.056180] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 765.056180] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] _ensure_no_port_binding_failure(port) [ 765.056180] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 765.056180] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] raise exception.PortBindingFailed(port_id=port['id']) [ 765.056180] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] nova.exception.PortBindingFailed: Binding failed for port 427f6f68-d9be-4108-b3aa-612f9786561c, please check neutron logs for more information. [ 765.056180] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] [ 765.056180] env[63021]: INFO nova.compute.manager [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] [instance: f77da305-9cc9-415d-a808-c2136212c456] Terminating instance [ 765.057070] env[63021]: DEBUG oslo_concurrency.lockutils [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Acquiring lock "refresh_cache-f77da305-9cc9-415d-a808-c2136212c456" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.060720] env[63021]: DEBUG nova.network.neutron [req-78cc699e-0fda-4ae2-93e4-fea3152a8364 req-116ff21a-395d-4472-b095-dd071e8f656d service nova] [instance: f77da305-9cc9-415d-a808-c2136212c456] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.338659] env[63021]: DEBUG oslo_concurrency.lockutils [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.380183] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.400s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.380701] env[63021]: DEBUG nova.compute.manager [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 765.383414] env[63021]: DEBUG oslo_concurrency.lockutils [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.563s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.384986] env[63021]: INFO nova.compute.claims [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 765.563242] env[63021]: DEBUG oslo_concurrency.lockutils [req-78cc699e-0fda-4ae2-93e4-fea3152a8364 req-116ff21a-395d-4472-b095-dd071e8f656d service nova] Releasing lock "refresh_cache-f77da305-9cc9-415d-a808-c2136212c456" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.563686] env[63021]: DEBUG oslo_concurrency.lockutils [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Acquired lock "refresh_cache-f77da305-9cc9-415d-a808-c2136212c456" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.563911] env[63021]: DEBUG nova.network.neutron [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] [instance: f77da305-9cc9-415d-a808-c2136212c456] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 765.889513] env[63021]: DEBUG nova.compute.utils [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 765.892919] env[63021]: DEBUG nova.compute.manager [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 765.893211] env[63021]: DEBUG nova.network.neutron [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 765.932144] env[63021]: DEBUG nova.policy [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ae2d6166369548b594a7214cf7e5d462', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c45c5b52db314a5587a009d560be94da', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 766.081730] env[63021]: DEBUG nova.network.neutron [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] [instance: f77da305-9cc9-415d-a808-c2136212c456] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.151978] env[63021]: DEBUG nova.network.neutron [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] [instance: f77da305-9cc9-415d-a808-c2136212c456] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.194433] env[63021]: DEBUG nova.network.neutron [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Successfully created port: 6dfc9d7b-a2ab-4aa4-990c-6f911a61b84b {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 766.393460] env[63021]: DEBUG nova.compute.manager [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 766.540427] env[63021]: DEBUG nova.compute.manager [req-03d6ed25-c017-4ed1-bf5c-3522b1cc064a req-26aa4955-3c41-4fe8-abc1-7784f625d57b service nova] [instance: f77da305-9cc9-415d-a808-c2136212c456] Received event network-vif-deleted-427f6f68-d9be-4108-b3aa-612f9786561c {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 766.655641] env[63021]: DEBUG oslo_concurrency.lockutils [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Releasing lock "refresh_cache-f77da305-9cc9-415d-a808-c2136212c456" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.656959] env[63021]: DEBUG nova.compute.manager [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] [instance: f77da305-9cc9-415d-a808-c2136212c456] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 766.656959] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] [instance: f77da305-9cc9-415d-a808-c2136212c456] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 766.656959] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-79089a83-a337-442e-87a1-7c2301861985 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.668017] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad5fba07-587f-40ef-9b75-5b23a6d27680 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.695217] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] [instance: f77da305-9cc9-415d-a808-c2136212c456] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f77da305-9cc9-415d-a808-c2136212c456 could not be found. [ 766.695217] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] [instance: f77da305-9cc9-415d-a808-c2136212c456] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 766.695217] env[63021]: INFO nova.compute.manager [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] [instance: f77da305-9cc9-415d-a808-c2136212c456] Took 0.04 seconds to destroy the instance on the hypervisor. [ 766.695217] env[63021]: DEBUG oslo.service.loopingcall [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 766.697720] env[63021]: DEBUG nova.compute.manager [-] [instance: f77da305-9cc9-415d-a808-c2136212c456] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 766.697897] env[63021]: DEBUG nova.network.neutron [-] [instance: f77da305-9cc9-415d-a808-c2136212c456] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 766.724741] env[63021]: DEBUG nova.network.neutron [-] [instance: f77da305-9cc9-415d-a808-c2136212c456] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.732590] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a72b55e6-0b3c-4374-b85c-a33ac4882439 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.740916] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44fe8309-1444-4cc1-813c-a2b72d648a22 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.770448] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb19963-d53f-4f88-b2a8-a18395b72e76 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.777987] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7fcb7d1-217b-4934-aa77-f53785b187f4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.791289] env[63021]: DEBUG nova.compute.provider_tree [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 767.059087] env[63021]: ERROR nova.compute.manager [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6dfc9d7b-a2ab-4aa4-990c-6f911a61b84b, please check neutron logs for more information. [ 767.059087] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 767.059087] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 767.059087] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 767.059087] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 767.059087] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 767.059087] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 767.059087] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 767.059087] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 767.059087] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 767.059087] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 767.059087] env[63021]: ERROR nova.compute.manager raise self.value [ 767.059087] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 767.059087] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 767.059087] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 767.059087] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 767.059876] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 767.059876] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 767.059876] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6dfc9d7b-a2ab-4aa4-990c-6f911a61b84b, please check neutron logs for more information. [ 767.059876] env[63021]: ERROR nova.compute.manager [ 767.059876] env[63021]: Traceback (most recent call last): [ 767.059876] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 767.059876] env[63021]: listener.cb(fileno) [ 767.059876] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 767.059876] env[63021]: result = function(*args, **kwargs) [ 767.059876] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 767.059876] env[63021]: return func(*args, **kwargs) [ 767.059876] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 767.059876] env[63021]: raise e [ 767.059876] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 767.059876] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 767.059876] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 767.059876] env[63021]: created_port_ids = self._update_ports_for_instance( [ 767.059876] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 767.059876] env[63021]: with excutils.save_and_reraise_exception(): [ 767.059876] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 767.059876] env[63021]: self.force_reraise() [ 767.059876] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 767.059876] env[63021]: raise self.value [ 767.059876] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 767.059876] env[63021]: updated_port = self._update_port( [ 767.059876] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 767.059876] env[63021]: _ensure_no_port_binding_failure(port) [ 767.059876] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 767.059876] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 767.060947] env[63021]: nova.exception.PortBindingFailed: Binding failed for port 6dfc9d7b-a2ab-4aa4-990c-6f911a61b84b, please check neutron logs for more information. [ 767.060947] env[63021]: Removing descriptor: 18 [ 767.227876] env[63021]: DEBUG nova.network.neutron [-] [instance: f77da305-9cc9-415d-a808-c2136212c456] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.294824] env[63021]: DEBUG nova.scheduler.client.report [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 767.409261] env[63021]: DEBUG nova.compute.manager [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 767.432581] env[63021]: DEBUG nova.virt.hardware [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 767.433202] env[63021]: DEBUG nova.virt.hardware [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 767.433202] env[63021]: DEBUG nova.virt.hardware [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 767.433337] env[63021]: DEBUG nova.virt.hardware [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 767.433474] env[63021]: DEBUG nova.virt.hardware [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 767.433618] env[63021]: DEBUG nova.virt.hardware [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 767.433821] env[63021]: DEBUG nova.virt.hardware [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 767.433985] env[63021]: DEBUG nova.virt.hardware [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 767.434201] env[63021]: DEBUG nova.virt.hardware [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 767.434368] env[63021]: DEBUG nova.virt.hardware [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 767.434534] env[63021]: DEBUG nova.virt.hardware [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 767.435377] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7892b2b9-2f36-4dd2-8fad-c3bacf91849e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.443068] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8ff79f7-1254-44ca-8f26-ee67cfbe103c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.457054] env[63021]: ERROR nova.compute.manager [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6dfc9d7b-a2ab-4aa4-990c-6f911a61b84b, please check neutron logs for more information. [ 767.457054] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Traceback (most recent call last): [ 767.457054] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 767.457054] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] yield resources [ 767.457054] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 767.457054] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] self.driver.spawn(context, instance, image_meta, [ 767.457054] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 767.457054] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] self._vmops.spawn(context, instance, image_meta, injected_files, [ 767.457054] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 767.457054] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] vm_ref = self.build_virtual_machine(instance, [ 767.457054] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 767.457505] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] vif_infos = vmwarevif.get_vif_info(self._session, [ 767.457505] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 767.457505] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] for vif in network_info: [ 767.457505] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 767.457505] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] return self._sync_wrapper(fn, *args, **kwargs) [ 767.457505] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 767.457505] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] self.wait() [ 767.457505] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 767.457505] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] self[:] = self._gt.wait() [ 767.457505] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 767.457505] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] return self._exit_event.wait() [ 767.457505] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 767.457505] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] current.throw(*self._exc) [ 767.457952] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 767.457952] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] result = function(*args, **kwargs) [ 767.457952] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 767.457952] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] return func(*args, **kwargs) [ 767.457952] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 767.457952] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] raise e [ 767.457952] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 767.457952] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] nwinfo = self.network_api.allocate_for_instance( [ 767.457952] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 767.457952] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] created_port_ids = self._update_ports_for_instance( [ 767.457952] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 767.457952] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] with excutils.save_and_reraise_exception(): [ 767.457952] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 767.458415] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] self.force_reraise() [ 767.458415] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 767.458415] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] raise self.value [ 767.458415] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 767.458415] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] updated_port = self._update_port( [ 767.458415] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 767.458415] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] _ensure_no_port_binding_failure(port) [ 767.458415] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 767.458415] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] raise exception.PortBindingFailed(port_id=port['id']) [ 767.458415] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] nova.exception.PortBindingFailed: Binding failed for port 6dfc9d7b-a2ab-4aa4-990c-6f911a61b84b, please check neutron logs for more information. [ 767.458415] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] [ 767.458415] env[63021]: INFO nova.compute.manager [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Terminating instance [ 767.459351] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Acquiring lock "refresh_cache-9f3b0e67-24fc-4625-81f4-050055906c19" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.459506] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Acquired lock "refresh_cache-9f3b0e67-24fc-4625-81f4-050055906c19" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.459665] env[63021]: DEBUG nova.network.neutron [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 767.729935] env[63021]: INFO nova.compute.manager [-] [instance: f77da305-9cc9-415d-a808-c2136212c456] Took 1.03 seconds to deallocate network for instance. [ 767.732447] env[63021]: DEBUG nova.compute.claims [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] [instance: f77da305-9cc9-415d-a808-c2136212c456] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 767.732653] env[63021]: DEBUG oslo_concurrency.lockutils [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.799711] env[63021]: DEBUG oslo_concurrency.lockutils [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.416s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.800321] env[63021]: DEBUG nova.compute.manager [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 767.802989] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.459s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.977226] env[63021]: DEBUG nova.network.neutron [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 768.048346] env[63021]: DEBUG nova.network.neutron [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.308806] env[63021]: DEBUG nova.compute.utils [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 768.313040] env[63021]: DEBUG nova.compute.manager [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 768.313040] env[63021]: DEBUG nova.network.neutron [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 768.350249] env[63021]: DEBUG nova.policy [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1e133aeda3249ec98dbdc53f869d719', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7331caed42034dfc81c79cfabbe0829e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 768.552353] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Releasing lock "refresh_cache-9f3b0e67-24fc-4625-81f4-050055906c19" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.552771] env[63021]: DEBUG nova.compute.manager [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 768.553025] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 768.553483] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fcffcb6e-4788-47e6-97b9-6b282ef3a866 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.562409] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9089767d-ea25-41ad-9f27-73bbd697f94d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.577725] env[63021]: DEBUG nova.compute.manager [req-3a5b468c-e8ff-4f9b-81a0-5551dd08a2ef req-348e7ac1-fd09-4e0e-ae02-f9a9e865ab9d service nova] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Received event network-changed-6dfc9d7b-a2ab-4aa4-990c-6f911a61b84b {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 768.577925] env[63021]: DEBUG nova.compute.manager [req-3a5b468c-e8ff-4f9b-81a0-5551dd08a2ef req-348e7ac1-fd09-4e0e-ae02-f9a9e865ab9d service nova] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Refreshing instance network info cache due to event network-changed-6dfc9d7b-a2ab-4aa4-990c-6f911a61b84b. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 768.578235] env[63021]: DEBUG oslo_concurrency.lockutils [req-3a5b468c-e8ff-4f9b-81a0-5551dd08a2ef req-348e7ac1-fd09-4e0e-ae02-f9a9e865ab9d service nova] Acquiring lock "refresh_cache-9f3b0e67-24fc-4625-81f4-050055906c19" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.578293] env[63021]: DEBUG oslo_concurrency.lockutils [req-3a5b468c-e8ff-4f9b-81a0-5551dd08a2ef req-348e7ac1-fd09-4e0e-ae02-f9a9e865ab9d service nova] Acquired lock "refresh_cache-9f3b0e67-24fc-4625-81f4-050055906c19" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.578444] env[63021]: DEBUG nova.network.neutron [req-3a5b468c-e8ff-4f9b-81a0-5551dd08a2ef req-348e7ac1-fd09-4e0e-ae02-f9a9e865ab9d service nova] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Refreshing network info cache for port 6dfc9d7b-a2ab-4aa4-990c-6f911a61b84b {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 768.592276] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9f3b0e67-24fc-4625-81f4-050055906c19 could not be found. [ 768.592276] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 768.592276] env[63021]: INFO nova.compute.manager [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Took 0.04 seconds to destroy the instance on the hypervisor. [ 768.592466] env[63021]: DEBUG oslo.service.loopingcall [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 768.594035] env[63021]: DEBUG nova.compute.manager [-] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 768.594144] env[63021]: DEBUG nova.network.neutron [-] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 768.596884] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-236645c9-e946-4af4-9d9b-271ea6afc907 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.605026] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a254017d-4510-4f2f-bfd7-c7385115a6a5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.610517] env[63021]: DEBUG nova.network.neutron [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Successfully created port: 95221085-b895-443b-ae1d-e9baca110bd3 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 768.636672] env[63021]: DEBUG nova.network.neutron [-] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 768.638821] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34071194-5839-4675-a690-4855c1a3a6c4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.647178] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce30402c-f9fb-4c92-b540-142a4db53aa5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.661046] env[63021]: DEBUG nova.compute.provider_tree [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 768.814377] env[63021]: DEBUG nova.compute.manager [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 769.095761] env[63021]: DEBUG nova.network.neutron [req-3a5b468c-e8ff-4f9b-81a0-5551dd08a2ef req-348e7ac1-fd09-4e0e-ae02-f9a9e865ab9d service nova] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 769.139484] env[63021]: DEBUG nova.network.neutron [-] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.164379] env[63021]: DEBUG nova.scheduler.client.report [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 769.187691] env[63021]: DEBUG nova.network.neutron [req-3a5b468c-e8ff-4f9b-81a0-5551dd08a2ef req-348e7ac1-fd09-4e0e-ae02-f9a9e865ab9d service nova] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.567382] env[63021]: ERROR nova.compute.manager [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 95221085-b895-443b-ae1d-e9baca110bd3, please check neutron logs for more information. [ 769.567382] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 769.567382] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.567382] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 769.567382] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 769.567382] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 769.567382] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 769.567382] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 769.567382] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.567382] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 769.567382] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.567382] env[63021]: ERROR nova.compute.manager raise self.value [ 769.567382] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 769.567382] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 769.567382] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.567382] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 769.568150] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.568150] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 769.568150] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 95221085-b895-443b-ae1d-e9baca110bd3, please check neutron logs for more information. [ 769.568150] env[63021]: ERROR nova.compute.manager [ 769.568150] env[63021]: Traceback (most recent call last): [ 769.568150] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 769.568150] env[63021]: listener.cb(fileno) [ 769.568150] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 769.568150] env[63021]: result = function(*args, **kwargs) [ 769.568150] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 769.568150] env[63021]: return func(*args, **kwargs) [ 769.568150] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 769.568150] env[63021]: raise e [ 769.568150] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.568150] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 769.568150] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 769.568150] env[63021]: created_port_ids = self._update_ports_for_instance( [ 769.568150] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 769.568150] env[63021]: with excutils.save_and_reraise_exception(): [ 769.568150] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.568150] env[63021]: self.force_reraise() [ 769.568150] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.568150] env[63021]: raise self.value [ 769.568150] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 769.568150] env[63021]: updated_port = self._update_port( [ 769.568150] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.568150] env[63021]: _ensure_no_port_binding_failure(port) [ 769.568150] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.568150] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 769.569108] env[63021]: nova.exception.PortBindingFailed: Binding failed for port 95221085-b895-443b-ae1d-e9baca110bd3, please check neutron logs for more information. [ 769.569108] env[63021]: Removing descriptor: 18 [ 769.644056] env[63021]: INFO nova.compute.manager [-] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Took 1.05 seconds to deallocate network for instance. [ 769.646643] env[63021]: DEBUG nova.compute.claims [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 769.646643] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.669959] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.866s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.670371] env[63021]: ERROR nova.compute.manager [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d88f0cc0-bc74-4b9e-842a-138776ad2023, please check neutron logs for more information. [ 769.670371] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Traceback (most recent call last): [ 769.670371] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 769.670371] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] self.driver.spawn(context, instance, image_meta, [ 769.670371] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 769.670371] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] self._vmops.spawn(context, instance, image_meta, injected_files, [ 769.670371] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 769.670371] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] vm_ref = self.build_virtual_machine(instance, [ 769.670371] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 769.670371] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] vif_infos = vmwarevif.get_vif_info(self._session, [ 769.670371] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 769.670777] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] for vif in network_info: [ 769.670777] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 769.670777] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] return self._sync_wrapper(fn, *args, **kwargs) [ 769.670777] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 769.670777] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] self.wait() [ 769.670777] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 769.670777] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] self[:] = self._gt.wait() [ 769.670777] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 769.670777] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] return self._exit_event.wait() [ 769.670777] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 769.670777] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] result = hub.switch() [ 769.670777] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 769.670777] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] return self.greenlet.switch() [ 769.671230] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 769.671230] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] result = function(*args, **kwargs) [ 769.671230] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 769.671230] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] return func(*args, **kwargs) [ 769.671230] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 769.671230] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] raise e [ 769.671230] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.671230] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] nwinfo = self.network_api.allocate_for_instance( [ 769.671230] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 769.671230] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] created_port_ids = self._update_ports_for_instance( [ 769.671230] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 769.671230] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] with excutils.save_and_reraise_exception(): [ 769.671230] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.671657] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] self.force_reraise() [ 769.671657] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.671657] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] raise self.value [ 769.671657] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 769.671657] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] updated_port = self._update_port( [ 769.671657] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.671657] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] _ensure_no_port_binding_failure(port) [ 769.671657] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.671657] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] raise exception.PortBindingFailed(port_id=port['id']) [ 769.671657] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] nova.exception.PortBindingFailed: Binding failed for port d88f0cc0-bc74-4b9e-842a-138776ad2023, please check neutron logs for more information. [ 769.671657] env[63021]: ERROR nova.compute.manager [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] [ 769.672086] env[63021]: DEBUG nova.compute.utils [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Binding failed for port d88f0cc0-bc74-4b9e-842a-138776ad2023, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 769.672687] env[63021]: DEBUG oslo_concurrency.lockutils [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.868s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.676019] env[63021]: DEBUG nova.compute.manager [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Build of instance 3660a926-7460-41f2-9ee4-2a5072701aea was re-scheduled: Binding failed for port d88f0cc0-bc74-4b9e-842a-138776ad2023, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 769.676473] env[63021]: DEBUG nova.compute.manager [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 769.676704] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Acquiring lock "refresh_cache-3660a926-7460-41f2-9ee4-2a5072701aea" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.676849] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Acquired lock "refresh_cache-3660a926-7460-41f2-9ee4-2a5072701aea" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.677011] env[63021]: DEBUG nova.network.neutron [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 769.689545] env[63021]: DEBUG oslo_concurrency.lockutils [req-3a5b468c-e8ff-4f9b-81a0-5551dd08a2ef req-348e7ac1-fd09-4e0e-ae02-f9a9e865ab9d service nova] Releasing lock "refresh_cache-9f3b0e67-24fc-4625-81f4-050055906c19" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.689935] env[63021]: DEBUG nova.compute.manager [req-3a5b468c-e8ff-4f9b-81a0-5551dd08a2ef req-348e7ac1-fd09-4e0e-ae02-f9a9e865ab9d service nova] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Received event network-vif-deleted-6dfc9d7b-a2ab-4aa4-990c-6f911a61b84b {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 769.823666] env[63021]: DEBUG nova.compute.manager [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 769.848793] env[63021]: DEBUG nova.virt.hardware [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 769.849041] env[63021]: DEBUG nova.virt.hardware [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 769.849198] env[63021]: DEBUG nova.virt.hardware [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 769.849372] env[63021]: DEBUG nova.virt.hardware [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 769.849514] env[63021]: DEBUG nova.virt.hardware [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 769.849655] env[63021]: DEBUG nova.virt.hardware [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 769.849856] env[63021]: DEBUG nova.virt.hardware [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 769.850014] env[63021]: DEBUG nova.virt.hardware [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 769.850180] env[63021]: DEBUG nova.virt.hardware [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 769.850340] env[63021]: DEBUG nova.virt.hardware [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 769.850508] env[63021]: DEBUG nova.virt.hardware [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 769.851691] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aca4e7d8-7e6a-455f-bb50-658be74a17f3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.859818] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3990d90-e005-4ee8-8667-e5db97a4445c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.873591] env[63021]: ERROR nova.compute.manager [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 95221085-b895-443b-ae1d-e9baca110bd3, please check neutron logs for more information. [ 769.873591] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Traceback (most recent call last): [ 769.873591] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 769.873591] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] yield resources [ 769.873591] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 769.873591] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] self.driver.spawn(context, instance, image_meta, [ 769.873591] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 769.873591] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 769.873591] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 769.873591] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] vm_ref = self.build_virtual_machine(instance, [ 769.873591] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 769.873997] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] vif_infos = vmwarevif.get_vif_info(self._session, [ 769.873997] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 769.873997] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] for vif in network_info: [ 769.873997] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 769.873997] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] return self._sync_wrapper(fn, *args, **kwargs) [ 769.873997] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 769.873997] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] self.wait() [ 769.873997] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 769.873997] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] self[:] = self._gt.wait() [ 769.873997] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 769.873997] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] return self._exit_event.wait() [ 769.873997] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 769.873997] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] current.throw(*self._exc) [ 769.874581] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 769.874581] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] result = function(*args, **kwargs) [ 769.874581] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 769.874581] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] return func(*args, **kwargs) [ 769.874581] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 769.874581] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] raise e [ 769.874581] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.874581] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] nwinfo = self.network_api.allocate_for_instance( [ 769.874581] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 769.874581] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] created_port_ids = self._update_ports_for_instance( [ 769.874581] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 769.874581] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] with excutils.save_and_reraise_exception(): [ 769.874581] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.875045] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] self.force_reraise() [ 769.875045] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.875045] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] raise self.value [ 769.875045] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 769.875045] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] updated_port = self._update_port( [ 769.875045] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.875045] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] _ensure_no_port_binding_failure(port) [ 769.875045] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.875045] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] raise exception.PortBindingFailed(port_id=port['id']) [ 769.875045] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] nova.exception.PortBindingFailed: Binding failed for port 95221085-b895-443b-ae1d-e9baca110bd3, please check neutron logs for more information. [ 769.875045] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] [ 769.875045] env[63021]: INFO nova.compute.manager [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Terminating instance [ 769.876291] env[63021]: DEBUG oslo_concurrency.lockutils [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "refresh_cache-ad2037c0-ed67-4cd1-b614-1733acfd7fd0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.876450] env[63021]: DEBUG oslo_concurrency.lockutils [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquired lock "refresh_cache-ad2037c0-ed67-4cd1-b614-1733acfd7fd0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.876611] env[63021]: DEBUG nova.network.neutron [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 770.198621] env[63021]: DEBUG nova.network.neutron [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 770.277089] env[63021]: DEBUG nova.network.neutron [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.393248] env[63021]: DEBUG nova.network.neutron [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 770.429248] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35689490-aef1-486b-9731-09f8aed67b8a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.436707] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-933a1ee4-6d03-4f93-b578-1decc843c15f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.465548] env[63021]: DEBUG nova.network.neutron [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.467040] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d66b9113-4076-40e6-92e7-7d7a420b7ec6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.473805] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bfe8829-41f1-40b1-9808-a743cb8bd736 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.488263] env[63021]: DEBUG nova.compute.provider_tree [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 770.590282] env[63021]: DEBUG nova.compute.manager [req-6c3c0b8c-a5ff-4494-8b32-131372f25f5d req-b89b2215-3477-4b98-8fd2-0b339e0a46de service nova] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Received event network-changed-95221085-b895-443b-ae1d-e9baca110bd3 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 770.590845] env[63021]: DEBUG nova.compute.manager [req-6c3c0b8c-a5ff-4494-8b32-131372f25f5d req-b89b2215-3477-4b98-8fd2-0b339e0a46de service nova] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Refreshing instance network info cache due to event network-changed-95221085-b895-443b-ae1d-e9baca110bd3. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 770.591086] env[63021]: DEBUG oslo_concurrency.lockutils [req-6c3c0b8c-a5ff-4494-8b32-131372f25f5d req-b89b2215-3477-4b98-8fd2-0b339e0a46de service nova] Acquiring lock "refresh_cache-ad2037c0-ed67-4cd1-b614-1733acfd7fd0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.781650] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Releasing lock "refresh_cache-3660a926-7460-41f2-9ee4-2a5072701aea" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.781911] env[63021]: DEBUG nova.compute.manager [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 770.782105] env[63021]: DEBUG nova.compute.manager [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 770.782272] env[63021]: DEBUG nova.network.neutron [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 770.798024] env[63021]: DEBUG nova.network.neutron [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 770.971300] env[63021]: DEBUG oslo_concurrency.lockutils [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Releasing lock "refresh_cache-ad2037c0-ed67-4cd1-b614-1733acfd7fd0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.971300] env[63021]: DEBUG nova.compute.manager [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 770.971559] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 770.971921] env[63021]: DEBUG oslo_concurrency.lockutils [req-6c3c0b8c-a5ff-4494-8b32-131372f25f5d req-b89b2215-3477-4b98-8fd2-0b339e0a46de service nova] Acquired lock "refresh_cache-ad2037c0-ed67-4cd1-b614-1733acfd7fd0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.972142] env[63021]: DEBUG nova.network.neutron [req-6c3c0b8c-a5ff-4494-8b32-131372f25f5d req-b89b2215-3477-4b98-8fd2-0b339e0a46de service nova] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Refreshing network info cache for port 95221085-b895-443b-ae1d-e9baca110bd3 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 770.973612] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fd0951c5-05b4-4ca0-a888-c42efa529a23 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.983810] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cec2b495-399f-4800-aa99-6baa824ca42e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.994504] env[63021]: DEBUG nova.scheduler.client.report [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 771.008160] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ad2037c0-ed67-4cd1-b614-1733acfd7fd0 could not be found. [ 771.008427] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 771.008640] env[63021]: INFO nova.compute.manager [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Took 0.04 seconds to destroy the instance on the hypervisor. [ 771.008902] env[63021]: DEBUG oslo.service.loopingcall [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 771.009209] env[63021]: DEBUG nova.compute.manager [-] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 771.009360] env[63021]: DEBUG nova.network.neutron [-] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 771.023741] env[63021]: DEBUG nova.network.neutron [-] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 771.300554] env[63021]: DEBUG nova.network.neutron [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.491829] env[63021]: DEBUG nova.network.neutron [req-6c3c0b8c-a5ff-4494-8b32-131372f25f5d req-b89b2215-3477-4b98-8fd2-0b339e0a46de service nova] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 771.499526] env[63021]: DEBUG oslo_concurrency.lockutils [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.827s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.500842] env[63021]: ERROR nova.compute.manager [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 09f2e8a1-f441-4cee-896f-7b33e52c6b5a, please check neutron logs for more information. [ 771.500842] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Traceback (most recent call last): [ 771.500842] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 771.500842] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] self.driver.spawn(context, instance, image_meta, [ 771.500842] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 771.500842] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 771.500842] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 771.500842] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] vm_ref = self.build_virtual_machine(instance, [ 771.500842] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 771.500842] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] vif_infos = vmwarevif.get_vif_info(self._session, [ 771.500842] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 771.501328] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] for vif in network_info: [ 771.501328] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 771.501328] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] return self._sync_wrapper(fn, *args, **kwargs) [ 771.501328] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 771.501328] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] self.wait() [ 771.501328] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 771.501328] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] self[:] = self._gt.wait() [ 771.501328] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 771.501328] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] return self._exit_event.wait() [ 771.501328] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 771.501328] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] result = hub.switch() [ 771.501328] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 771.501328] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] return self.greenlet.switch() [ 771.501758] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 771.501758] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] result = function(*args, **kwargs) [ 771.501758] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 771.501758] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] return func(*args, **kwargs) [ 771.501758] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 771.501758] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] raise e [ 771.501758] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 771.501758] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] nwinfo = self.network_api.allocate_for_instance( [ 771.501758] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 771.501758] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] created_port_ids = self._update_ports_for_instance( [ 771.501758] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 771.501758] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] with excutils.save_and_reraise_exception(): [ 771.501758] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 771.502237] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] self.force_reraise() [ 771.502237] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 771.502237] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] raise self.value [ 771.502237] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 771.502237] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] updated_port = self._update_port( [ 771.502237] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 771.502237] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] _ensure_no_port_binding_failure(port) [ 771.502237] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 771.502237] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] raise exception.PortBindingFailed(port_id=port['id']) [ 771.502237] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] nova.exception.PortBindingFailed: Binding failed for port 09f2e8a1-f441-4cee-896f-7b33e52c6b5a, please check neutron logs for more information. [ 771.502237] env[63021]: ERROR nova.compute.manager [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] [ 771.502565] env[63021]: DEBUG nova.compute.utils [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Binding failed for port 09f2e8a1-f441-4cee-896f-7b33e52c6b5a, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 771.502565] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.604s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.504973] env[63021]: DEBUG nova.compute.manager [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Build of instance bc48ac80-36f6-4c97-9f95-29aed2efeae6 was re-scheduled: Binding failed for port 09f2e8a1-f441-4cee-896f-7b33e52c6b5a, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 771.505455] env[63021]: DEBUG nova.compute.manager [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 771.505681] env[63021]: DEBUG oslo_concurrency.lockutils [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Acquiring lock "refresh_cache-bc48ac80-36f6-4c97-9f95-29aed2efeae6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.505825] env[63021]: DEBUG oslo_concurrency.lockutils [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Acquired lock "refresh_cache-bc48ac80-36f6-4c97-9f95-29aed2efeae6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.505982] env[63021]: DEBUG nova.network.neutron [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 771.525937] env[63021]: DEBUG nova.network.neutron [-] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.563218] env[63021]: DEBUG nova.network.neutron [req-6c3c0b8c-a5ff-4494-8b32-131372f25f5d req-b89b2215-3477-4b98-8fd2-0b339e0a46de service nova] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.803239] env[63021]: INFO nova.compute.manager [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] [instance: 3660a926-7460-41f2-9ee4-2a5072701aea] Took 1.02 seconds to deallocate network for instance. [ 772.028757] env[63021]: DEBUG nova.network.neutron [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 772.030586] env[63021]: INFO nova.compute.manager [-] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Took 1.02 seconds to deallocate network for instance. [ 772.035889] env[63021]: DEBUG nova.compute.claims [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 772.035889] env[63021]: DEBUG oslo_concurrency.lockutils [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.069126] env[63021]: DEBUG oslo_concurrency.lockutils [req-6c3c0b8c-a5ff-4494-8b32-131372f25f5d req-b89b2215-3477-4b98-8fd2-0b339e0a46de service nova] Releasing lock "refresh_cache-ad2037c0-ed67-4cd1-b614-1733acfd7fd0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.069394] env[63021]: DEBUG nova.compute.manager [req-6c3c0b8c-a5ff-4494-8b32-131372f25f5d req-b89b2215-3477-4b98-8fd2-0b339e0a46de service nova] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Received event network-vif-deleted-95221085-b895-443b-ae1d-e9baca110bd3 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 772.107183] env[63021]: DEBUG nova.network.neutron [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.292979] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a019f991-69c3-4b12-89fb-8e60188dba39 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.300296] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-938f01bb-278e-415f-bd64-4fa7e91f8f1e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.332738] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-584718d1-f197-4b1e-b51d-ed49df73c595 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.340421] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-646eb98c-8cf4-437f-a5b8-ed4fbb4720f5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.353620] env[63021]: DEBUG nova.compute.provider_tree [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 772.609917] env[63021]: DEBUG oslo_concurrency.lockutils [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Releasing lock "refresh_cache-bc48ac80-36f6-4c97-9f95-29aed2efeae6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.610190] env[63021]: DEBUG nova.compute.manager [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 772.610387] env[63021]: DEBUG nova.compute.manager [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 772.610555] env[63021]: DEBUG nova.network.neutron [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 772.626145] env[63021]: DEBUG nova.network.neutron [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 772.856127] env[63021]: DEBUG nova.scheduler.client.report [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 772.860922] env[63021]: INFO nova.scheduler.client.report [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Deleted allocations for instance 3660a926-7460-41f2-9ee4-2a5072701aea [ 773.128825] env[63021]: DEBUG nova.network.neutron [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.368465] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.866s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.369017] env[63021]: ERROR nova.compute.manager [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5cea81bb-fd5c-430e-b8f1-5013d52a6802, please check neutron logs for more information. [ 773.369017] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Traceback (most recent call last): [ 773.369017] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 773.369017] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] self.driver.spawn(context, instance, image_meta, [ 773.369017] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 773.369017] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] self._vmops.spawn(context, instance, image_meta, injected_files, [ 773.369017] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 773.369017] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] vm_ref = self.build_virtual_machine(instance, [ 773.369017] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 773.369017] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] vif_infos = vmwarevif.get_vif_info(self._session, [ 773.369017] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 773.369408] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] for vif in network_info: [ 773.369408] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 773.369408] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] return self._sync_wrapper(fn, *args, **kwargs) [ 773.369408] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 773.369408] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] self.wait() [ 773.369408] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 773.369408] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] self[:] = self._gt.wait() [ 773.369408] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 773.369408] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] return self._exit_event.wait() [ 773.369408] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 773.369408] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] current.throw(*self._exc) [ 773.369408] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 773.369408] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] result = function(*args, **kwargs) [ 773.369839] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 773.369839] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] return func(*args, **kwargs) [ 773.369839] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 773.369839] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] raise e [ 773.369839] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 773.369839] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] nwinfo = self.network_api.allocate_for_instance( [ 773.369839] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 773.369839] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] created_port_ids = self._update_ports_for_instance( [ 773.369839] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 773.369839] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] with excutils.save_and_reraise_exception(): [ 773.369839] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 773.369839] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] self.force_reraise() [ 773.369839] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 773.370303] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] raise self.value [ 773.370303] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 773.370303] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] updated_port = self._update_port( [ 773.370303] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 773.370303] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] _ensure_no_port_binding_failure(port) [ 773.370303] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 773.370303] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] raise exception.PortBindingFailed(port_id=port['id']) [ 773.370303] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] nova.exception.PortBindingFailed: Binding failed for port 5cea81bb-fd5c-430e-b8f1-5013d52a6802, please check neutron logs for more information. [ 773.370303] env[63021]: ERROR nova.compute.manager [instance: 907540da-a701-477a-9b5c-9942b5d2d987] [ 773.370303] env[63021]: DEBUG nova.compute.utils [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Binding failed for port 5cea81bb-fd5c-430e-b8f1-5013d52a6802, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 773.370980] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.262s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.375059] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7e57d6f6-0fdc-4cc4-a107-dc314a781c85 tempest-ListServerFiltersTestJSON-174169111 tempest-ListServerFiltersTestJSON-174169111-project-member] Lock "3660a926-7460-41f2-9ee4-2a5072701aea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 139.088s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.375532] env[63021]: DEBUG nova.compute.manager [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Build of instance 907540da-a701-477a-9b5c-9942b5d2d987 was re-scheduled: Binding failed for port 5cea81bb-fd5c-430e-b8f1-5013d52a6802, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 773.375938] env[63021]: DEBUG nova.compute.manager [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 773.376171] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "refresh_cache-907540da-a701-477a-9b5c-9942b5d2d987" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.376319] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquired lock "refresh_cache-907540da-a701-477a-9b5c-9942b5d2d987" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.376474] env[63021]: DEBUG nova.network.neutron [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 773.631938] env[63021]: INFO nova.compute.manager [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] [instance: bc48ac80-36f6-4c97-9f95-29aed2efeae6] Took 1.02 seconds to deallocate network for instance. [ 773.883475] env[63021]: DEBUG nova.compute.manager [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 773.899871] env[63021]: DEBUG nova.network.neutron [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 773.994995] env[63021]: DEBUG nova.network.neutron [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.164525] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61502f34-9cda-4da9-b440-cf6139c2f770 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.172600] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70dd9436-2642-4056-9ef6-8d22633ff981 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.203670] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b29a4a28-0bb0-4dc6-954b-efb31bcf79d7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.211733] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d77a058e-fa08-4da7-a6e3-40161776fbf7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.225679] env[63021]: DEBUG nova.compute.provider_tree [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 774.403037] env[63021]: DEBUG oslo_concurrency.lockutils [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.498231] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Releasing lock "refresh_cache-907540da-a701-477a-9b5c-9942b5d2d987" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.498474] env[63021]: DEBUG nova.compute.manager [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 774.498654] env[63021]: DEBUG nova.compute.manager [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 774.498816] env[63021]: DEBUG nova.network.neutron [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 774.513637] env[63021]: DEBUG nova.network.neutron [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 774.665657] env[63021]: INFO nova.scheduler.client.report [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Deleted allocations for instance bc48ac80-36f6-4c97-9f95-29aed2efeae6 [ 774.728872] env[63021]: DEBUG nova.scheduler.client.report [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 775.016502] env[63021]: DEBUG nova.network.neutron [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.174206] env[63021]: DEBUG oslo_concurrency.lockutils [None req-a9067ce6-c72e-4d06-b437-c172211bc426 tempest-ServersNegativeTestJSON-1391883409 tempest-ServersNegativeTestJSON-1391883409-project-member] Lock "bc48ac80-36f6-4c97-9f95-29aed2efeae6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 135.244s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.234244] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.863s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.234843] env[63021]: ERROR nova.compute.manager [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c4616f5a-0905-4083-83b8-33fb1d8dd324, please check neutron logs for more information. [ 775.234843] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Traceback (most recent call last): [ 775.234843] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 775.234843] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] self.driver.spawn(context, instance, image_meta, [ 775.234843] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 775.234843] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 775.234843] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 775.234843] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] vm_ref = self.build_virtual_machine(instance, [ 775.234843] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 775.234843] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] vif_infos = vmwarevif.get_vif_info(self._session, [ 775.234843] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 775.235251] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] for vif in network_info: [ 775.235251] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 775.235251] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] return self._sync_wrapper(fn, *args, **kwargs) [ 775.235251] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 775.235251] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] self.wait() [ 775.235251] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 775.235251] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] self[:] = self._gt.wait() [ 775.235251] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 775.235251] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] return self._exit_event.wait() [ 775.235251] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 775.235251] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] result = hub.switch() [ 775.235251] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 775.235251] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] return self.greenlet.switch() [ 775.235735] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 775.235735] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] result = function(*args, **kwargs) [ 775.235735] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 775.235735] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] return func(*args, **kwargs) [ 775.235735] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 775.235735] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] raise e [ 775.235735] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 775.235735] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] nwinfo = self.network_api.allocate_for_instance( [ 775.235735] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 775.235735] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] created_port_ids = self._update_ports_for_instance( [ 775.235735] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 775.235735] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] with excutils.save_and_reraise_exception(): [ 775.235735] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 775.236193] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] self.force_reraise() [ 775.236193] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 775.236193] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] raise self.value [ 775.236193] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 775.236193] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] updated_port = self._update_port( [ 775.236193] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 775.236193] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] _ensure_no_port_binding_failure(port) [ 775.236193] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 775.236193] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] raise exception.PortBindingFailed(port_id=port['id']) [ 775.236193] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] nova.exception.PortBindingFailed: Binding failed for port c4616f5a-0905-4083-83b8-33fb1d8dd324, please check neutron logs for more information. [ 775.236193] env[63021]: ERROR nova.compute.manager [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] [ 775.236608] env[63021]: DEBUG nova.compute.utils [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Binding failed for port c4616f5a-0905-4083-83b8-33fb1d8dd324, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 775.237232] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.603s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.240719] env[63021]: DEBUG nova.compute.manager [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Build of instance 06ef77f1-a196-499c-b639-8427b021a8c0 was re-scheduled: Binding failed for port c4616f5a-0905-4083-83b8-33fb1d8dd324, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 775.241446] env[63021]: DEBUG nova.compute.manager [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 775.241729] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Acquiring lock "refresh_cache-06ef77f1-a196-499c-b639-8427b021a8c0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.241891] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Acquired lock "refresh_cache-06ef77f1-a196-499c-b639-8427b021a8c0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.242065] env[63021]: DEBUG nova.network.neutron [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 775.519728] env[63021]: INFO nova.compute.manager [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 907540da-a701-477a-9b5c-9942b5d2d987] Took 1.02 seconds to deallocate network for instance. [ 775.677273] env[63021]: DEBUG nova.compute.manager [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 775.771728] env[63021]: DEBUG nova.network.neutron [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 775.972334] env[63021]: DEBUG nova.network.neutron [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.059185] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28c02699-2d40-45b3-b1b0-8464b1813305 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.069799] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d6e83af-8367-4c1e-8c35-3569496e6970 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.104153] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1c21416-6444-4871-9c9d-8e9a82b69b4c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.114052] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55d46195-4e48-4899-89ee-4c2384f28bf5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.128294] env[63021]: DEBUG nova.compute.provider_tree [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.197763] env[63021]: DEBUG oslo_concurrency.lockutils [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.474691] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Releasing lock "refresh_cache-06ef77f1-a196-499c-b639-8427b021a8c0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.474938] env[63021]: DEBUG nova.compute.manager [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 776.475142] env[63021]: DEBUG nova.compute.manager [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 776.475338] env[63021]: DEBUG nova.network.neutron [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 776.498716] env[63021]: DEBUG nova.network.neutron [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 776.560097] env[63021]: INFO nova.scheduler.client.report [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Deleted allocations for instance 907540da-a701-477a-9b5c-9942b5d2d987 [ 776.631544] env[63021]: DEBUG nova.scheduler.client.report [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 777.001994] env[63021]: DEBUG nova.network.neutron [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.073661] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1903add1-bfb8-49db-8da6-3a9fd550aec4 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "907540da-a701-477a-9b5c-9942b5d2d987" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.302s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.136466] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.899s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.137226] env[63021]: ERROR nova.compute.manager [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 08a404ab-ec2a-4859-b302-f163e69680c2, please check neutron logs for more information. [ 777.137226] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Traceback (most recent call last): [ 777.137226] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 777.137226] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] self.driver.spawn(context, instance, image_meta, [ 777.137226] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 777.137226] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 777.137226] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 777.137226] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] vm_ref = self.build_virtual_machine(instance, [ 777.137226] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 777.137226] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] vif_infos = vmwarevif.get_vif_info(self._session, [ 777.137226] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 777.137961] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] for vif in network_info: [ 777.137961] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 777.137961] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] return self._sync_wrapper(fn, *args, **kwargs) [ 777.137961] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 777.137961] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] self.wait() [ 777.137961] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 777.137961] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] self[:] = self._gt.wait() [ 777.137961] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 777.137961] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] return self._exit_event.wait() [ 777.137961] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 777.137961] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] result = hub.switch() [ 777.137961] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 777.137961] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] return self.greenlet.switch() [ 777.138654] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 777.138654] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] result = function(*args, **kwargs) [ 777.138654] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 777.138654] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] return func(*args, **kwargs) [ 777.138654] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 777.138654] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] raise e [ 777.138654] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 777.138654] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] nwinfo = self.network_api.allocate_for_instance( [ 777.138654] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 777.138654] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] created_port_ids = self._update_ports_for_instance( [ 777.138654] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 777.138654] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] with excutils.save_and_reraise_exception(): [ 777.138654] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 777.139355] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] self.force_reraise() [ 777.139355] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 777.139355] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] raise self.value [ 777.139355] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 777.139355] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] updated_port = self._update_port( [ 777.139355] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 777.139355] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] _ensure_no_port_binding_failure(port) [ 777.139355] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 777.139355] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] raise exception.PortBindingFailed(port_id=port['id']) [ 777.139355] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] nova.exception.PortBindingFailed: Binding failed for port 08a404ab-ec2a-4859-b302-f163e69680c2, please check neutron logs for more information. [ 777.139355] env[63021]: ERROR nova.compute.manager [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] [ 777.139961] env[63021]: DEBUG nova.compute.utils [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Binding failed for port 08a404ab-ec2a-4859-b302-f163e69680c2, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 777.139961] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 14.094s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.141486] env[63021]: DEBUG nova.compute.manager [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Build of instance 8ea2a693-0cd8-4f91-8bd8-b40265b9372e was re-scheduled: Binding failed for port 08a404ab-ec2a-4859-b302-f163e69680c2, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 777.141914] env[63021]: DEBUG nova.compute.manager [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 777.142240] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Acquiring lock "refresh_cache-8ea2a693-0cd8-4f91-8bd8-b40265b9372e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.142350] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Acquired lock "refresh_cache-8ea2a693-0cd8-4f91-8bd8-b40265b9372e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.142460] env[63021]: DEBUG nova.network.neutron [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 777.508836] env[63021]: INFO nova.compute.manager [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 06ef77f1-a196-499c-b639-8427b021a8c0] Took 1.03 seconds to deallocate network for instance. [ 777.582698] env[63021]: DEBUG nova.compute.manager [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 777.670373] env[63021]: DEBUG nova.network.neutron [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 777.755338] env[63021]: DEBUG nova.network.neutron [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.114868] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.260515] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Releasing lock "refresh_cache-8ea2a693-0cd8-4f91-8bd8-b40265b9372e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.260515] env[63021]: DEBUG nova.compute.manager [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 778.260515] env[63021]: DEBUG nova.compute.manager [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 778.260515] env[63021]: DEBUG nova.network.neutron [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 778.287906] env[63021]: DEBUG nova.network.neutron [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 778.550899] env[63021]: INFO nova.scheduler.client.report [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Deleted allocations for instance 06ef77f1-a196-499c-b639-8427b021a8c0 [ 778.692828] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 06ef77f1-a196-499c-b639-8427b021a8c0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 778.792154] env[63021]: DEBUG nova.network.neutron [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.073622] env[63021]: DEBUG oslo_concurrency.lockutils [None req-7311ea65-c330-4a5b-9116-92164465c559 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lock "06ef77f1-a196-499c-b639-8427b021a8c0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 138.193s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.196745] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 8ea2a693-0cd8-4f91-8bd8-b40265b9372e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 779.197165] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance d02adae0-724e-4dbe-a6c5-d476805d3d57 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 779.197165] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance f77da305-9cc9-415d-a808-c2136212c456 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 779.197165] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 9f3b0e67-24fc-4625-81f4-050055906c19 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 779.197316] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance ad2037c0-ed67-4cd1-b614-1733acfd7fd0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 779.207244] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "71909213-258d-45f5-9a3e-e473e8ce8aa2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.207244] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "71909213-258d-45f5-9a3e-e473e8ce8aa2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.294453] env[63021]: INFO nova.compute.manager [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] [instance: 8ea2a693-0cd8-4f91-8bd8-b40265b9372e] Took 1.03 seconds to deallocate network for instance. [ 779.577653] env[63021]: DEBUG nova.compute.manager [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 779.703961] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance f225812d-f6bf-4350-9151-9ab6c54cce5d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 780.102619] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.207691] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 9a6720ef-5307-44b7-b566-8d553c9b5384 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 780.329097] env[63021]: INFO nova.scheduler.client.report [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Deleted allocations for instance 8ea2a693-0cd8-4f91-8bd8-b40265b9372e [ 780.707463] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquiring lock "df856146-61ba-4de7-a5fd-d75263927163" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.707703] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "df856146-61ba-4de7-a5fd-d75263927163" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.710604] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 929cf4cd-6f3c-4988-8f80-d889d3070e1e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 780.840312] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6425db9d-f3b6-477b-89f1-84e9826daaf1 tempest-InstanceActionsV221TestJSON-1175767528 tempest-InstanceActionsV221TestJSON-1175767528-project-member] Lock "8ea2a693-0cd8-4f91-8bd8-b40265b9372e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 135.104s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.213920] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 7c6d7839-616d-41f5-a909-f8c626477a5e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 781.344154] env[63021]: DEBUG nova.compute.manager [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 781.718402] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance bef5c7cb-be2f-4be7-b72c-4a066d473dd0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 781.866039] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.168821] env[63021]: DEBUG oslo_concurrency.lockutils [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Acquiring lock "45c9e2be-eb19-4b83-b280-c9eeaddfccba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.169098] env[63021]: DEBUG oslo_concurrency.lockutils [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lock "45c9e2be-eb19-4b83-b280-c9eeaddfccba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.221513] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 36ec45ec-761b-4d62-a74f-e4d9a840ada0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 782.724446] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 868b1aeb-8109-4cb9-bb58-33f3d92989ad has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 783.228307] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance df0b6974-7f78-4b73-8583-d09754fbacea has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 783.732571] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 2209d709-7a5b-42e2-be93-b4fd436b6290 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 784.234642] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 36d26354-6127-4557-acb0-8743e27ff1ff has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 784.738324] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 785.244868] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 6f4e6245-c0d9-46b4-baf5-5aca19e66da5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 785.750757] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 70f67905-7c60-433d-9ebc-d66fa44eb36e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 786.251869] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 7787f6bb-e855-476f-9146-2a648932429d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 786.253049] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=63021) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 786.253049] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=63021) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 786.508499] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92418c70-4036-43a6-9285-6a81ab0ed036 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.516511] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2a71714-ba32-4358-989c-9d62f4e24518 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.554586] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f025854-da1f-45a6-8830-44a358118cd3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.562132] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e41f11d5-c8e4-45cd-b83d-f3f0ce32d452 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.576236] env[63021]: DEBUG nova.compute.provider_tree [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 787.084645] env[63021]: DEBUG nova.scheduler.client.report [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 787.150070] env[63021]: DEBUG oslo_concurrency.lockutils [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "c328d2f7-3398-4f25-b11c-f464be7af8a1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.150306] env[63021]: DEBUG oslo_concurrency.lockutils [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "c328d2f7-3398-4f25-b11c-f464be7af8a1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.589057] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63021) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 787.589383] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 10.450s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.589523] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.033s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.599076] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 787.599273] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Cleaning up deleted instances {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 788.105498] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] There are 4 instances to clean {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 788.105873] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 77fd1199-33b5-4040-a0ba-dd881da1224c] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 788.432638] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c3a5d92-9721-41a5-92e7-f73ed86e5253 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.441107] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00f598a1-844a-4044-9be0-e143b7baba11 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.470732] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9edbfc0-a3d2-4f1d-8860-a0bf07be2363 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.477705] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec977f1c-bcc8-452c-852b-a2e0676c35db {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.490091] env[63021]: DEBUG nova.compute.provider_tree [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 788.612499] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 9d474bd0-e825-49f2-9715-14d85495718e] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 788.992871] env[63021]: DEBUG nova.scheduler.client.report [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 789.115657] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 5ca0039c-b9c8-4ee0-bbe5-e0f68494dadd] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 789.498086] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.908s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.498734] env[63021]: ERROR nova.compute.manager [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6cab39b0-f517-48fa-916b-2a14b0be3861, please check neutron logs for more information. [ 789.498734] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Traceback (most recent call last): [ 789.498734] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 789.498734] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] self.driver.spawn(context, instance, image_meta, [ 789.498734] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 789.498734] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] self._vmops.spawn(context, instance, image_meta, injected_files, [ 789.498734] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 789.498734] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] vm_ref = self.build_virtual_machine(instance, [ 789.498734] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 789.498734] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] vif_infos = vmwarevif.get_vif_info(self._session, [ 789.498734] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 789.499107] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] for vif in network_info: [ 789.499107] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 789.499107] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] return self._sync_wrapper(fn, *args, **kwargs) [ 789.499107] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 789.499107] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] self.wait() [ 789.499107] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 789.499107] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] self[:] = self._gt.wait() [ 789.499107] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 789.499107] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] return self._exit_event.wait() [ 789.499107] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 789.499107] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] result = hub.switch() [ 789.499107] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 789.499107] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] return self.greenlet.switch() [ 789.499532] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 789.499532] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] result = function(*args, **kwargs) [ 789.499532] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 789.499532] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] return func(*args, **kwargs) [ 789.499532] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 789.499532] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] raise e [ 789.499532] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 789.499532] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] nwinfo = self.network_api.allocate_for_instance( [ 789.499532] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 789.499532] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] created_port_ids = self._update_ports_for_instance( [ 789.499532] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 789.499532] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] with excutils.save_and_reraise_exception(): [ 789.499532] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 789.499968] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] self.force_reraise() [ 789.499968] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 789.499968] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] raise self.value [ 789.499968] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 789.499968] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] updated_port = self._update_port( [ 789.499968] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 789.499968] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] _ensure_no_port_binding_failure(port) [ 789.499968] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 789.499968] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] raise exception.PortBindingFailed(port_id=port['id']) [ 789.499968] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] nova.exception.PortBindingFailed: Binding failed for port 6cab39b0-f517-48fa-916b-2a14b0be3861, please check neutron logs for more information. [ 789.499968] env[63021]: ERROR nova.compute.manager [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] [ 789.500351] env[63021]: DEBUG nova.compute.utils [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Binding failed for port 6cab39b0-f517-48fa-916b-2a14b0be3861, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 789.500720] env[63021]: DEBUG oslo_concurrency.lockutils [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.162s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.502213] env[63021]: INFO nova.compute.claims [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 789.504838] env[63021]: DEBUG nova.compute.manager [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Build of instance d02adae0-724e-4dbe-a6c5-d476805d3d57 was re-scheduled: Binding failed for port 6cab39b0-f517-48fa-916b-2a14b0be3861, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 789.505271] env[63021]: DEBUG nova.compute.manager [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 789.505497] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Acquiring lock "refresh_cache-d02adae0-724e-4dbe-a6c5-d476805d3d57" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.505643] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Acquired lock "refresh_cache-d02adae0-724e-4dbe-a6c5-d476805d3d57" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.505797] env[63021]: DEBUG nova.network.neutron [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 789.619645] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 0950cbd3-0708-4994-8d2d-053aecc06cdf] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 790.026461] env[63021]: DEBUG nova.network.neutron [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 790.111556] env[63021]: DEBUG nova.network.neutron [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.123376] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 790.123376] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Cleaning up deleted instances with incomplete migration {{(pid=63021) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 790.614558] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Releasing lock "refresh_cache-d02adae0-724e-4dbe-a6c5-d476805d3d57" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.614796] env[63021]: DEBUG nova.compute.manager [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 790.614982] env[63021]: DEBUG nova.compute.manager [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 790.615161] env[63021]: DEBUG nova.network.neutron [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 790.625574] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 790.630857] env[63021]: DEBUG nova.network.neutron [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 790.776593] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98224374-1490-4cb6-b4e7-b704fe7f4fbb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.784261] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de07dc07-d201-45ec-87a7-cdf92b6f14a3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.814427] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4576ef0-7a33-4d56-a9d4-a520a6ba0cf0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.821243] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f8fd5a1-1386-485a-a8f6-1286d6b79286 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.833817] env[63021]: DEBUG nova.compute.provider_tree [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 791.133088] env[63021]: DEBUG nova.network.neutron [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.336427] env[63021]: DEBUG nova.scheduler.client.report [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 791.637301] env[63021]: INFO nova.compute.manager [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] [instance: d02adae0-724e-4dbe-a6c5-d476805d3d57] Took 1.02 seconds to deallocate network for instance. [ 791.844400] env[63021]: DEBUG oslo_concurrency.lockutils [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.344s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.844943] env[63021]: DEBUG nova.compute.manager [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 791.847641] env[63021]: DEBUG oslo_concurrency.lockutils [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.115s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.351550] env[63021]: DEBUG nova.compute.utils [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 792.356349] env[63021]: DEBUG nova.compute.manager [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 792.356951] env[63021]: DEBUG nova.network.neutron [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 792.397082] env[63021]: DEBUG nova.policy [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f37ac2ae39ec43669736e1de99adb298', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f64fe0a46b07403f9ce4317b8a743121', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 792.634029] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37b18997-a2a3-4b30-a2e1-b3bc3c93eee5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.642062] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e8691d1-7d5d-466a-baab-e666779612dd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.681943] env[63021]: DEBUG nova.network.neutron [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Successfully created port: d62d2305-e5cd-4344-a697-444a752049d9 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 792.684371] env[63021]: INFO nova.scheduler.client.report [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Deleted allocations for instance d02adae0-724e-4dbe-a6c5-d476805d3d57 [ 792.690307] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d1e7822-7218-4cde-9606-4fa819938daa {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.698879] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6888f658-e86b-4729-949d-64df4f33fff7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.712846] env[63021]: DEBUG nova.compute.provider_tree [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 792.857055] env[63021]: DEBUG nova.compute.manager [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 793.123165] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 793.123454] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 793.123605] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Starting heal instance info cache {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 793.123727] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Rebuilding the list of instances to heal {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 793.196447] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6d7f9ba7-6caf-4019-a88f-771298c0766d tempest-ServersTestMultiNic-1224203138 tempest-ServersTestMultiNic-1224203138-project-member] Lock "d02adae0-724e-4dbe-a6c5-d476805d3d57" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.975s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.218017] env[63021]: DEBUG nova.scheduler.client.report [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 793.347119] env[63021]: DEBUG nova.compute.manager [req-917ef4e4-93c1-43f3-8f48-1e3bcb9c5382 req-ce10f864-7d80-428d-b8a7-4fae12a8523f service nova] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Received event network-changed-d62d2305-e5cd-4344-a697-444a752049d9 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 793.347588] env[63021]: DEBUG nova.compute.manager [req-917ef4e4-93c1-43f3-8f48-1e3bcb9c5382 req-ce10f864-7d80-428d-b8a7-4fae12a8523f service nova] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Refreshing instance network info cache due to event network-changed-d62d2305-e5cd-4344-a697-444a752049d9. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 793.349935] env[63021]: DEBUG oslo_concurrency.lockutils [req-917ef4e4-93c1-43f3-8f48-1e3bcb9c5382 req-ce10f864-7d80-428d-b8a7-4fae12a8523f service nova] Acquiring lock "refresh_cache-f225812d-f6bf-4350-9151-9ab6c54cce5d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.349935] env[63021]: DEBUG oslo_concurrency.lockutils [req-917ef4e4-93c1-43f3-8f48-1e3bcb9c5382 req-ce10f864-7d80-428d-b8a7-4fae12a8523f service nova] Acquired lock "refresh_cache-f225812d-f6bf-4350-9151-9ab6c54cce5d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.349935] env[63021]: DEBUG nova.network.neutron [req-917ef4e4-93c1-43f3-8f48-1e3bcb9c5382 req-ce10f864-7d80-428d-b8a7-4fae12a8523f service nova] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Refreshing network info cache for port d62d2305-e5cd-4344-a697-444a752049d9 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 793.523195] env[63021]: ERROR nova.compute.manager [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d62d2305-e5cd-4344-a697-444a752049d9, please check neutron logs for more information. [ 793.523195] env[63021]: ERROR nova.compute.manager Traceback (most recent call last): [ 793.523195] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 793.523195] env[63021]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 793.523195] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 793.523195] env[63021]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 793.523195] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 793.523195] env[63021]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 793.523195] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 793.523195] env[63021]: ERROR nova.compute.manager self.force_reraise() [ 793.523195] env[63021]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 793.523195] env[63021]: ERROR nova.compute.manager raise self.value [ 793.523195] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 793.523195] env[63021]: ERROR nova.compute.manager updated_port = self._update_port( [ 793.523195] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 793.523195] env[63021]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 793.524016] env[63021]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 793.524016] env[63021]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 793.524016] env[63021]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d62d2305-e5cd-4344-a697-444a752049d9, please check neutron logs for more information. [ 793.524016] env[63021]: ERROR nova.compute.manager [ 793.524016] env[63021]: Traceback (most recent call last): [ 793.524016] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 793.524016] env[63021]: listener.cb(fileno) [ 793.524016] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 793.524016] env[63021]: result = function(*args, **kwargs) [ 793.524016] env[63021]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 793.524016] env[63021]: return func(*args, **kwargs) [ 793.524016] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 793.524016] env[63021]: raise e [ 793.524016] env[63021]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 793.524016] env[63021]: nwinfo = self.network_api.allocate_for_instance( [ 793.524016] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 793.524016] env[63021]: created_port_ids = self._update_ports_for_instance( [ 793.524016] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 793.524016] env[63021]: with excutils.save_and_reraise_exception(): [ 793.524016] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 793.524016] env[63021]: self.force_reraise() [ 793.524016] env[63021]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 793.524016] env[63021]: raise self.value [ 793.524016] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 793.524016] env[63021]: updated_port = self._update_port( [ 793.524016] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 793.524016] env[63021]: _ensure_no_port_binding_failure(port) [ 793.524016] env[63021]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 793.524016] env[63021]: raise exception.PortBindingFailed(port_id=port['id']) [ 793.524979] env[63021]: nova.exception.PortBindingFailed: Binding failed for port d62d2305-e5cd-4344-a697-444a752049d9, please check neutron logs for more information. [ 793.524979] env[63021]: Removing descriptor: 18 [ 793.628696] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Skipping network cache update for instance because it is Building. {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 793.628884] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Skipping network cache update for instance because it is Building. {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 793.629033] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Skipping network cache update for instance because it is Building. {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 793.629166] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Didn't find any instances for network info cache update. {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 793.629367] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 793.629521] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 793.629662] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 793.629809] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 793.629948] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 793.630301] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 793.630301] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63021) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 793.630416] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 793.698971] env[63021]: DEBUG nova.compute.manager [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 793.720854] env[63021]: DEBUG oslo_concurrency.lockutils [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.873s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.721538] env[63021]: ERROR nova.compute.manager [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] [instance: f77da305-9cc9-415d-a808-c2136212c456] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 427f6f68-d9be-4108-b3aa-612f9786561c, please check neutron logs for more information. [ 793.721538] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] Traceback (most recent call last): [ 793.721538] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 793.721538] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] self.driver.spawn(context, instance, image_meta, [ 793.721538] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 793.721538] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] self._vmops.spawn(context, instance, image_meta, injected_files, [ 793.721538] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 793.721538] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] vm_ref = self.build_virtual_machine(instance, [ 793.721538] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 793.721538] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] vif_infos = vmwarevif.get_vif_info(self._session, [ 793.721538] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 793.721914] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] for vif in network_info: [ 793.721914] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 793.721914] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] return self._sync_wrapper(fn, *args, **kwargs) [ 793.721914] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 793.721914] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] self.wait() [ 793.721914] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 793.721914] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] self[:] = self._gt.wait() [ 793.721914] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 793.721914] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] return self._exit_event.wait() [ 793.721914] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 793.721914] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] current.throw(*self._exc) [ 793.721914] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 793.721914] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] result = function(*args, **kwargs) [ 793.722321] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 793.722321] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] return func(*args, **kwargs) [ 793.722321] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 793.722321] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] raise e [ 793.722321] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 793.722321] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] nwinfo = self.network_api.allocate_for_instance( [ 793.722321] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 793.722321] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] created_port_ids = self._update_ports_for_instance( [ 793.722321] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 793.722321] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] with excutils.save_and_reraise_exception(): [ 793.722321] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 793.722321] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] self.force_reraise() [ 793.722321] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 793.722701] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] raise self.value [ 793.722701] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 793.722701] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] updated_port = self._update_port( [ 793.722701] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 793.722701] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] _ensure_no_port_binding_failure(port) [ 793.722701] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 793.722701] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] raise exception.PortBindingFailed(port_id=port['id']) [ 793.722701] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] nova.exception.PortBindingFailed: Binding failed for port 427f6f68-d9be-4108-b3aa-612f9786561c, please check neutron logs for more information. [ 793.722701] env[63021]: ERROR nova.compute.manager [instance: f77da305-9cc9-415d-a808-c2136212c456] [ 793.722701] env[63021]: DEBUG nova.compute.utils [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] [instance: f77da305-9cc9-415d-a808-c2136212c456] Binding failed for port 427f6f68-d9be-4108-b3aa-612f9786561c, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 793.725287] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.079s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.734087] env[63021]: DEBUG nova.compute.manager [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] [instance: f77da305-9cc9-415d-a808-c2136212c456] Build of instance f77da305-9cc9-415d-a808-c2136212c456 was re-scheduled: Binding failed for port 427f6f68-d9be-4108-b3aa-612f9786561c, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 793.735286] env[63021]: DEBUG nova.compute.manager [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] [instance: f77da305-9cc9-415d-a808-c2136212c456] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 793.735620] env[63021]: DEBUG oslo_concurrency.lockutils [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Acquiring lock "refresh_cache-f77da305-9cc9-415d-a808-c2136212c456" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.735852] env[63021]: DEBUG oslo_concurrency.lockutils [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Acquired lock "refresh_cache-f77da305-9cc9-415d-a808-c2136212c456" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.736076] env[63021]: DEBUG nova.network.neutron [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] [instance: f77da305-9cc9-415d-a808-c2136212c456] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 793.867234] env[63021]: DEBUG nova.compute.manager [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 793.873390] env[63021]: DEBUG nova.network.neutron [req-917ef4e4-93c1-43f3-8f48-1e3bcb9c5382 req-ce10f864-7d80-428d-b8a7-4fae12a8523f service nova] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 793.894575] env[63021]: DEBUG nova.virt.hardware [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 793.894824] env[63021]: DEBUG nova.virt.hardware [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 793.894976] env[63021]: DEBUG nova.virt.hardware [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 793.895177] env[63021]: DEBUG nova.virt.hardware [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 793.895321] env[63021]: DEBUG nova.virt.hardware [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 793.895465] env[63021]: DEBUG nova.virt.hardware [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 793.895698] env[63021]: DEBUG nova.virt.hardware [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 793.895861] env[63021]: DEBUG nova.virt.hardware [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 793.896048] env[63021]: DEBUG nova.virt.hardware [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 793.896214] env[63021]: DEBUG nova.virt.hardware [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 793.896383] env[63021]: DEBUG nova.virt.hardware [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 793.897258] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ad730e0-78b6-4025-b2df-fc2462e6de10 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.905399] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5085788-80eb-4f89-9e20-efdd3dce6761 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.918849] env[63021]: ERROR nova.compute.manager [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d62d2305-e5cd-4344-a697-444a752049d9, please check neutron logs for more information. [ 793.918849] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Traceback (most recent call last): [ 793.918849] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 793.918849] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] yield resources [ 793.918849] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 793.918849] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] self.driver.spawn(context, instance, image_meta, [ 793.918849] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 793.918849] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 793.918849] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 793.918849] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] vm_ref = self.build_virtual_machine(instance, [ 793.918849] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 793.919294] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] vif_infos = vmwarevif.get_vif_info(self._session, [ 793.919294] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 793.919294] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] for vif in network_info: [ 793.919294] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 793.919294] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] return self._sync_wrapper(fn, *args, **kwargs) [ 793.919294] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 793.919294] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] self.wait() [ 793.919294] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 793.919294] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] self[:] = self._gt.wait() [ 793.919294] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 793.919294] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] return self._exit_event.wait() [ 793.919294] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 793.919294] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] current.throw(*self._exc) [ 793.919757] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 793.919757] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] result = function(*args, **kwargs) [ 793.919757] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 793.919757] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] return func(*args, **kwargs) [ 793.919757] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 793.919757] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] raise e [ 793.919757] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 793.919757] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] nwinfo = self.network_api.allocate_for_instance( [ 793.919757] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 793.919757] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] created_port_ids = self._update_ports_for_instance( [ 793.919757] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 793.919757] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] with excutils.save_and_reraise_exception(): [ 793.919757] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 793.920190] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] self.force_reraise() [ 793.920190] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 793.920190] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] raise self.value [ 793.920190] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 793.920190] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] updated_port = self._update_port( [ 793.920190] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 793.920190] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] _ensure_no_port_binding_failure(port) [ 793.920190] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 793.920190] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] raise exception.PortBindingFailed(port_id=port['id']) [ 793.920190] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] nova.exception.PortBindingFailed: Binding failed for port d62d2305-e5cd-4344-a697-444a752049d9, please check neutron logs for more information. [ 793.920190] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] [ 793.920190] env[63021]: INFO nova.compute.manager [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Terminating instance [ 793.921417] env[63021]: DEBUG oslo_concurrency.lockutils [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Acquiring lock "refresh_cache-f225812d-f6bf-4350-9151-9ab6c54cce5d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.009703] env[63021]: DEBUG nova.network.neutron [req-917ef4e4-93c1-43f3-8f48-1e3bcb9c5382 req-ce10f864-7d80-428d-b8a7-4fae12a8523f service nova] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.132801] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.225362] env[63021]: DEBUG oslo_concurrency.lockutils [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.260037] env[63021]: DEBUG nova.network.neutron [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] [instance: f77da305-9cc9-415d-a808-c2136212c456] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 794.334183] env[63021]: DEBUG nova.network.neutron [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] [instance: f77da305-9cc9-415d-a808-c2136212c456] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.506393] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a1c4dd0-55c7-42ef-91a8-42640e0e0ddb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.513095] env[63021]: DEBUG oslo_concurrency.lockutils [req-917ef4e4-93c1-43f3-8f48-1e3bcb9c5382 req-ce10f864-7d80-428d-b8a7-4fae12a8523f service nova] Releasing lock "refresh_cache-f225812d-f6bf-4350-9151-9ab6c54cce5d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.513584] env[63021]: DEBUG oslo_concurrency.lockutils [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Acquired lock "refresh_cache-f225812d-f6bf-4350-9151-9ab6c54cce5d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.513755] env[63021]: DEBUG nova.network.neutron [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 794.515460] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6da630e5-bde6-4d03-a49b-3d3e8096a6e0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.545350] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b27845a-0eb2-4183-a071-2b682a655031 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.556017] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3d47a52-af47-4fe7-bc07-f8fb224aad28 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.567075] env[63021]: DEBUG nova.compute.provider_tree [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 794.838324] env[63021]: DEBUG oslo_concurrency.lockutils [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Releasing lock "refresh_cache-f77da305-9cc9-415d-a808-c2136212c456" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.838608] env[63021]: DEBUG nova.compute.manager [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 794.838748] env[63021]: DEBUG nova.compute.manager [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] [instance: f77da305-9cc9-415d-a808-c2136212c456] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 794.838909] env[63021]: DEBUG nova.network.neutron [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] [instance: f77da305-9cc9-415d-a808-c2136212c456] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 794.855550] env[63021]: DEBUG nova.network.neutron [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] [instance: f77da305-9cc9-415d-a808-c2136212c456] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 795.037489] env[63021]: DEBUG nova.network.neutron [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 795.069501] env[63021]: DEBUG nova.scheduler.client.report [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 795.113781] env[63021]: DEBUG nova.network.neutron [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.357881] env[63021]: DEBUG nova.network.neutron [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] [instance: f77da305-9cc9-415d-a808-c2136212c456] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.370186] env[63021]: DEBUG nova.compute.manager [req-d1e5d653-5d24-4f1a-916d-e6d9d78cd8ba req-3250499f-dc34-4992-99eb-12d6542e41df service nova] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Received event network-vif-deleted-d62d2305-e5cd-4344-a697-444a752049d9 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 795.575149] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.850s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.575818] env[63021]: ERROR nova.compute.manager [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6dfc9d7b-a2ab-4aa4-990c-6f911a61b84b, please check neutron logs for more information. [ 795.575818] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Traceback (most recent call last): [ 795.575818] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 795.575818] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] self.driver.spawn(context, instance, image_meta, [ 795.575818] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 795.575818] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] self._vmops.spawn(context, instance, image_meta, injected_files, [ 795.575818] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 795.575818] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] vm_ref = self.build_virtual_machine(instance, [ 795.575818] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 795.575818] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] vif_infos = vmwarevif.get_vif_info(self._session, [ 795.575818] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 795.576258] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] for vif in network_info: [ 795.576258] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 795.576258] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] return self._sync_wrapper(fn, *args, **kwargs) [ 795.576258] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 795.576258] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] self.wait() [ 795.576258] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 795.576258] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] self[:] = self._gt.wait() [ 795.576258] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 795.576258] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] return self._exit_event.wait() [ 795.576258] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 795.576258] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] current.throw(*self._exc) [ 795.576258] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 795.576258] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] result = function(*args, **kwargs) [ 795.576644] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 795.576644] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] return func(*args, **kwargs) [ 795.576644] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 795.576644] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] raise e [ 795.576644] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 795.576644] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] nwinfo = self.network_api.allocate_for_instance( [ 795.576644] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 795.576644] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] created_port_ids = self._update_ports_for_instance( [ 795.576644] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 795.576644] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] with excutils.save_and_reraise_exception(): [ 795.576644] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 795.576644] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] self.force_reraise() [ 795.576644] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 795.577043] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] raise self.value [ 795.577043] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 795.577043] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] updated_port = self._update_port( [ 795.577043] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 795.577043] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] _ensure_no_port_binding_failure(port) [ 795.577043] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 795.577043] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] raise exception.PortBindingFailed(port_id=port['id']) [ 795.577043] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] nova.exception.PortBindingFailed: Binding failed for port 6dfc9d7b-a2ab-4aa4-990c-6f911a61b84b, please check neutron logs for more information. [ 795.577043] env[63021]: ERROR nova.compute.manager [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] [ 795.577043] env[63021]: DEBUG nova.compute.utils [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Binding failed for port 6dfc9d7b-a2ab-4aa4-990c-6f911a61b84b, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 795.577895] env[63021]: DEBUG oslo_concurrency.lockutils [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.542s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.581200] env[63021]: DEBUG nova.compute.manager [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Build of instance 9f3b0e67-24fc-4625-81f4-050055906c19 was re-scheduled: Binding failed for port 6dfc9d7b-a2ab-4aa4-990c-6f911a61b84b, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 795.581474] env[63021]: DEBUG nova.compute.manager [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 795.581698] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Acquiring lock "refresh_cache-9f3b0e67-24fc-4625-81f4-050055906c19" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.582570] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Acquired lock "refresh_cache-9f3b0e67-24fc-4625-81f4-050055906c19" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.582570] env[63021]: DEBUG nova.network.neutron [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 795.616518] env[63021]: DEBUG oslo_concurrency.lockutils [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Releasing lock "refresh_cache-f225812d-f6bf-4350-9151-9ab6c54cce5d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.616933] env[63021]: DEBUG nova.compute.manager [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 795.617137] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 795.617673] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a87aa5f3-0cb0-4d1f-8e37-e56aef62b701 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.627365] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9b2c283-96e7-49ec-bfe6-38874b1b1692 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.649030] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f225812d-f6bf-4350-9151-9ab6c54cce5d could not be found. [ 795.649309] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 795.649494] env[63021]: INFO nova.compute.manager [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Took 0.03 seconds to destroy the instance on the hypervisor. [ 795.649730] env[63021]: DEBUG oslo.service.loopingcall [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 795.649954] env[63021]: DEBUG nova.compute.manager [-] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 795.650056] env[63021]: DEBUG nova.network.neutron [-] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 795.665800] env[63021]: DEBUG nova.network.neutron [-] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 795.861768] env[63021]: INFO nova.compute.manager [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] [instance: f77da305-9cc9-415d-a808-c2136212c456] Took 1.02 seconds to deallocate network for instance. [ 796.104827] env[63021]: DEBUG nova.network.neutron [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 796.167969] env[63021]: DEBUG nova.network.neutron [-] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.210381] env[63021]: DEBUG nova.network.neutron [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.357727] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99296c52-4a77-453a-b498-d530de048b59 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.365025] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbd772b8-cfd4-4083-bb56-a32e93133e48 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.398843] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ae3b34a-92ae-4775-9051-f88b1201c0ba {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.407762] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76034e1f-91a6-41c2-8a8f-78f43cf48d9c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.421638] env[63021]: DEBUG nova.compute.provider_tree [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 796.670421] env[63021]: INFO nova.compute.manager [-] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Took 1.02 seconds to deallocate network for instance. [ 796.673745] env[63021]: DEBUG nova.compute.claims [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Aborting claim: {{(pid=63021) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 796.673918] env[63021]: DEBUG oslo_concurrency.lockutils [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.713855] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Releasing lock "refresh_cache-9f3b0e67-24fc-4625-81f4-050055906c19" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.714140] env[63021]: DEBUG nova.compute.manager [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 796.714321] env[63021]: DEBUG nova.compute.manager [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 796.714478] env[63021]: DEBUG nova.network.neutron [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 796.730113] env[63021]: DEBUG nova.network.neutron [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 796.902864] env[63021]: INFO nova.scheduler.client.report [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Deleted allocations for instance f77da305-9cc9-415d-a808-c2136212c456 [ 796.926586] env[63021]: DEBUG nova.scheduler.client.report [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 797.233120] env[63021]: DEBUG nova.network.neutron [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.412729] env[63021]: DEBUG oslo_concurrency.lockutils [None req-712fb6cd-1d9c-46eb-9375-1d0ee4f9fd5b tempest-ServerPasswordTestJSON-981918268 tempest-ServerPasswordTestJSON-981918268-project-member] Lock "f77da305-9cc9-415d-a808-c2136212c456" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.266s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.433932] env[63021]: DEBUG oslo_concurrency.lockutils [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.856s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.434616] env[63021]: ERROR nova.compute.manager [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 95221085-b895-443b-ae1d-e9baca110bd3, please check neutron logs for more information. [ 797.434616] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Traceback (most recent call last): [ 797.434616] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 797.434616] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] self.driver.spawn(context, instance, image_meta, [ 797.434616] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 797.434616] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 797.434616] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 797.434616] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] vm_ref = self.build_virtual_machine(instance, [ 797.434616] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 797.434616] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] vif_infos = vmwarevif.get_vif_info(self._session, [ 797.434616] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 797.435033] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] for vif in network_info: [ 797.435033] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 797.435033] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] return self._sync_wrapper(fn, *args, **kwargs) [ 797.435033] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 797.435033] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] self.wait() [ 797.435033] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 797.435033] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] self[:] = self._gt.wait() [ 797.435033] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 797.435033] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] return self._exit_event.wait() [ 797.435033] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 797.435033] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] current.throw(*self._exc) [ 797.435033] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 797.435033] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] result = function(*args, **kwargs) [ 797.435442] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 797.435442] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] return func(*args, **kwargs) [ 797.435442] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 797.435442] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] raise e [ 797.435442] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 797.435442] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] nwinfo = self.network_api.allocate_for_instance( [ 797.435442] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 797.435442] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] created_port_ids = self._update_ports_for_instance( [ 797.435442] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 797.435442] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] with excutils.save_and_reraise_exception(): [ 797.435442] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 797.435442] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] self.force_reraise() [ 797.435442] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 797.436094] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] raise self.value [ 797.436094] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 797.436094] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] updated_port = self._update_port( [ 797.436094] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 797.436094] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] _ensure_no_port_binding_failure(port) [ 797.436094] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 797.436094] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] raise exception.PortBindingFailed(port_id=port['id']) [ 797.436094] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] nova.exception.PortBindingFailed: Binding failed for port 95221085-b895-443b-ae1d-e9baca110bd3, please check neutron logs for more information. [ 797.436094] env[63021]: ERROR nova.compute.manager [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] [ 797.436094] env[63021]: DEBUG nova.compute.utils [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Binding failed for port 95221085-b895-443b-ae1d-e9baca110bd3, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 797.436755] env[63021]: DEBUG oslo_concurrency.lockutils [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.034s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.439869] env[63021]: INFO nova.compute.claims [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 797.442735] env[63021]: DEBUG nova.compute.manager [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Build of instance ad2037c0-ed67-4cd1-b614-1733acfd7fd0 was re-scheduled: Binding failed for port 95221085-b895-443b-ae1d-e9baca110bd3, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 797.443168] env[63021]: DEBUG nova.compute.manager [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 797.443419] env[63021]: DEBUG oslo_concurrency.lockutils [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "refresh_cache-ad2037c0-ed67-4cd1-b614-1733acfd7fd0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.443574] env[63021]: DEBUG oslo_concurrency.lockutils [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquired lock "refresh_cache-ad2037c0-ed67-4cd1-b614-1733acfd7fd0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.443729] env[63021]: DEBUG nova.network.neutron [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 797.738418] env[63021]: INFO nova.compute.manager [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] [instance: 9f3b0e67-24fc-4625-81f4-050055906c19] Took 1.02 seconds to deallocate network for instance. [ 797.916103] env[63021]: DEBUG nova.compute.manager [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 797.969294] env[63021]: DEBUG nova.network.neutron [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 798.058883] env[63021]: DEBUG nova.network.neutron [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.564135] env[63021]: DEBUG oslo_concurrency.lockutils [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Releasing lock "refresh_cache-ad2037c0-ed67-4cd1-b614-1733acfd7fd0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.564135] env[63021]: DEBUG nova.compute.manager [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 798.564135] env[63021]: DEBUG nova.compute.manager [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 798.564135] env[63021]: DEBUG nova.network.neutron [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 798.565132] env[63021]: DEBUG oslo_concurrency.lockutils [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.594819] env[63021]: DEBUG nova.network.neutron [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 798.777679] env[63021]: INFO nova.scheduler.client.report [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Deleted allocations for instance 9f3b0e67-24fc-4625-81f4-050055906c19 [ 798.872288] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e749debd-f2b0-4ce0-aea9-652760be9a1b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.881788] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-982d874f-cdd8-4c35-a1f7-2632e2938068 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.918180] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0fe0651-78eb-491d-92bc-7c1f3a9a50c9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.925988] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a280b488-9738-49f0-9833-8b92d9d73363 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.939281] env[63021]: DEBUG nova.compute.provider_tree [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 799.101346] env[63021]: DEBUG nova.network.neutron [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.290785] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6be2d8da-2ff8-4545-b1ec-47bfdffab98a tempest-ServersTestManualDisk-833406318 tempest-ServersTestManualDisk-833406318-project-member] Lock "9f3b0e67-24fc-4625-81f4-050055906c19" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 130.712s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.443957] env[63021]: DEBUG nova.scheduler.client.report [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 799.603767] env[63021]: INFO nova.compute.manager [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: ad2037c0-ed67-4cd1-b614-1733acfd7fd0] Took 1.04 seconds to deallocate network for instance. [ 799.794474] env[63021]: DEBUG nova.compute.manager [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 799.952460] env[63021]: DEBUG oslo_concurrency.lockutils [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.515s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.952876] env[63021]: DEBUG nova.compute.manager [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 799.959568] env[63021]: DEBUG oslo_concurrency.lockutils [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.759s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.959568] env[63021]: INFO nova.compute.claims [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 800.326487] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.466112] env[63021]: DEBUG nova.compute.utils [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 800.469227] env[63021]: DEBUG nova.compute.manager [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Not allocating networking since 'none' was specified. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 800.647229] env[63021]: INFO nova.scheduler.client.report [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Deleted allocations for instance ad2037c0-ed67-4cd1-b614-1733acfd7fd0 [ 800.653441] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "16b21798-17e8-4aeb-affa-57eae31d5dd3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.653712] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "16b21798-17e8-4aeb-affa-57eae31d5dd3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.972765] env[63021]: DEBUG nova.compute.manager [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 801.157286] env[63021]: DEBUG oslo_concurrency.lockutils [None req-20ba2932-b823-445d-b6bf-6d011bf86101 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "ad2037c0-ed67-4cd1-b614-1733acfd7fd0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 126.541s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.259582] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc2ca9f2-41f5-4019-ab3b-6757248ec4ed {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.266989] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e826c5a-15a4-4b5d-9c38-9c0a160d12ec {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.296303] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77813bf1-2b77-4937-a7d1-61c0daf6b87d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.303480] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01180109-0c37-491c-9c3d-45ab0d8f0aa9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.316719] env[63021]: DEBUG nova.compute.provider_tree [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 801.659860] env[63021]: DEBUG nova.compute.manager [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 801.819332] env[63021]: DEBUG nova.scheduler.client.report [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 801.989313] env[63021]: DEBUG nova.compute.manager [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 802.012137] env[63021]: DEBUG nova.virt.hardware [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 802.012648] env[63021]: DEBUG nova.virt.hardware [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 802.012842] env[63021]: DEBUG nova.virt.hardware [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 802.013213] env[63021]: DEBUG nova.virt.hardware [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 802.013432] env[63021]: DEBUG nova.virt.hardware [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 802.013591] env[63021]: DEBUG nova.virt.hardware [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 802.013823] env[63021]: DEBUG nova.virt.hardware [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 802.013993] env[63021]: DEBUG nova.virt.hardware [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 802.014193] env[63021]: DEBUG nova.virt.hardware [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 802.014599] env[63021]: DEBUG nova.virt.hardware [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 802.014785] env[63021]: DEBUG nova.virt.hardware [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 802.015755] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adca0d92-f76f-420c-b633-2559ced5a81a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.024161] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10c6eaf9-4967-41fd-9e6a-c1bcaa3350fd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.038282] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Instance VIF info [] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 802.044860] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Creating folder: Project (60293ae01cd0449d9f4c7be520167395). Parent ref: group-v277447. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 802.045151] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4c8b5cde-d893-4d78-8642-66ad5d081dde {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.054052] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Created folder: Project (60293ae01cd0449d9f4c7be520167395) in parent group-v277447. [ 802.054242] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Creating folder: Instances. Parent ref: group-v277463. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 802.054484] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-df791059-0945-414f-8fcf-8fd12d98e22e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.062440] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Created folder: Instances in parent group-v277463. [ 802.062687] env[63021]: DEBUG oslo.service.loopingcall [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 802.062872] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 802.063092] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9dac6d9f-b4ed-4762-9acd-225c486514c1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.078193] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 802.078193] env[63021]: value = "task-1293593" [ 802.078193] env[63021]: _type = "Task" [ 802.078193] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.088102] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293593, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.182142] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.324940] env[63021]: DEBUG oslo_concurrency.lockutils [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.368s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.325564] env[63021]: DEBUG nova.compute.manager [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 802.328316] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.214s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.329820] env[63021]: INFO nova.compute.claims [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 802.589048] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293593, 'name': CreateVM_Task, 'duration_secs': 0.301539} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.589801] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 802.590197] env[63021]: DEBUG oslo_concurrency.lockutils [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.590400] env[63021]: DEBUG oslo_concurrency.lockutils [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.590717] env[63021]: DEBUG oslo_concurrency.lockutils [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 802.590954] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6072e040-2c2a-492a-8c58-04fa8f6addb7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.595476] env[63021]: DEBUG oslo_vmware.api [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Waiting for the task: (returnval){ [ 802.595476] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52102bb1-8aa1-6b4b-8a6a-c22d9fb4d0d2" [ 802.595476] env[63021]: _type = "Task" [ 802.595476] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.602856] env[63021]: DEBUG oslo_vmware.api [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52102bb1-8aa1-6b4b-8a6a-c22d9fb4d0d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.834783] env[63021]: DEBUG nova.compute.utils [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 802.838904] env[63021]: DEBUG nova.compute.manager [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 802.839644] env[63021]: DEBUG nova.network.neutron [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 802.906559] env[63021]: DEBUG nova.policy [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c6ae7dd068ee473a8adbc86cca8cb1b2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5f1a621cb616420bb530cc905d86948b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 802.962997] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "881c2fb8-5f8c-48c2-a173-8fab949f24f7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.963365] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "881c2fb8-5f8c-48c2-a173-8fab949f24f7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.107854] env[63021]: DEBUG oslo_vmware.api [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52102bb1-8aa1-6b4b-8a6a-c22d9fb4d0d2, 'name': SearchDatastore_Task, 'duration_secs': 0.01043} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.109359] env[63021]: DEBUG oslo_concurrency.lockutils [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.109359] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 803.109359] env[63021]: DEBUG oslo_concurrency.lockutils [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.109359] env[63021]: DEBUG oslo_concurrency.lockutils [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.109550] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 803.109550] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0f7b0f6e-34dd-42ec-8aa5-0543f2e234b1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.117370] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 803.117559] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 803.118368] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8a6a03c-b33b-4767-b5ee-d85b1f821540 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.124116] env[63021]: DEBUG oslo_vmware.api [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Waiting for the task: (returnval){ [ 803.124116] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52eeeeb4-a7c4-8ae0-0533-5f311ed27640" [ 803.124116] env[63021]: _type = "Task" [ 803.124116] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.131916] env[63021]: DEBUG oslo_vmware.api [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52eeeeb4-a7c4-8ae0-0533-5f311ed27640, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.222411] env[63021]: DEBUG nova.network.neutron [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Successfully created port: 034fe798-4091-44e1-9dcd-618125f3a581 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 803.343314] env[63021]: DEBUG nova.compute.manager [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 803.636397] env[63021]: DEBUG oslo_vmware.api [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52eeeeb4-a7c4-8ae0-0533-5f311ed27640, 'name': SearchDatastore_Task, 'duration_secs': 0.012298} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.640332] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c914314-df2d-45aa-9f0c-05257619e0a7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.647027] env[63021]: DEBUG oslo_vmware.api [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Waiting for the task: (returnval){ [ 803.647027] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52264fcc-8c8d-5138-b28f-24a19ac2ded8" [ 803.647027] env[63021]: _type = "Task" [ 803.647027] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.659753] env[63021]: DEBUG oslo_vmware.api [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52264fcc-8c8d-5138-b28f-24a19ac2ded8, 'name': SearchDatastore_Task, 'duration_secs': 0.010274} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.660042] env[63021]: DEBUG oslo_concurrency.lockutils [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.660304] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 9a6720ef-5307-44b7-b566-8d553c9b5384/9a6720ef-5307-44b7-b566-8d553c9b5384.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 803.662940] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5c79fea4-7326-41d3-99f1-e152e234ba35 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.674645] env[63021]: DEBUG oslo_vmware.api [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Waiting for the task: (returnval){ [ 803.674645] env[63021]: value = "task-1293594" [ 803.674645] env[63021]: _type = "Task" [ 803.674645] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.682807] env[63021]: DEBUG oslo_vmware.api [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293594, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.684598] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a369a70-36bf-4169-81b9-950721f42346 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.691684] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5cf9f9c-5753-402b-99ba-a6a24b59276d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.723956] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a4b411d-12c1-4062-b755-2ef7b09e15cc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.730459] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65fcb10c-98fe-4a75-8bde-f8d5cc7dd8a8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.749959] env[63021]: DEBUG nova.compute.provider_tree [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 804.185751] env[63021]: DEBUG oslo_vmware.api [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293594, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.253618] env[63021]: DEBUG nova.scheduler.client.report [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 804.355969] env[63021]: DEBUG nova.compute.manager [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 804.379992] env[63021]: DEBUG nova.virt.hardware [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 804.380339] env[63021]: DEBUG nova.virt.hardware [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 804.380549] env[63021]: DEBUG nova.virt.hardware [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 804.380801] env[63021]: DEBUG nova.virt.hardware [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 804.380999] env[63021]: DEBUG nova.virt.hardware [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 804.381221] env[63021]: DEBUG nova.virt.hardware [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 804.381499] env[63021]: DEBUG nova.virt.hardware [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 804.381718] env[63021]: DEBUG nova.virt.hardware [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 804.381946] env[63021]: DEBUG nova.virt.hardware [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 804.382184] env[63021]: DEBUG nova.virt.hardware [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 804.382419] env[63021]: DEBUG nova.virt.hardware [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 804.383637] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7b2619b-b389-42f2-85a4-8f324bd0c3e5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.393895] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c690415e-a678-4035-99b5-672570a2d2e9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.685686] env[63021]: DEBUG oslo_vmware.api [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293594, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.56952} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.685964] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 9a6720ef-5307-44b7-b566-8d553c9b5384/9a6720ef-5307-44b7-b566-8d553c9b5384.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 804.686202] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 804.686461] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-208b87f4-c13d-4f1b-94ef-fd180b6b5ca8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.694048] env[63021]: DEBUG oslo_vmware.api [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Waiting for the task: (returnval){ [ 804.694048] env[63021]: value = "task-1293595" [ 804.694048] env[63021]: _type = "Task" [ 804.694048] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.702044] env[63021]: DEBUG oslo_vmware.api [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293595, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.746060] env[63021]: DEBUG nova.compute.manager [req-0ae252cb-a68e-47d8-b724-b369bac0bec7 req-efb0dd87-8116-423f-8217-de31fa87fd38 service nova] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Received event network-vif-plugged-034fe798-4091-44e1-9dcd-618125f3a581 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 804.746286] env[63021]: DEBUG oslo_concurrency.lockutils [req-0ae252cb-a68e-47d8-b724-b369bac0bec7 req-efb0dd87-8116-423f-8217-de31fa87fd38 service nova] Acquiring lock "929cf4cd-6f3c-4988-8f80-d889d3070e1e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.746489] env[63021]: DEBUG oslo_concurrency.lockutils [req-0ae252cb-a68e-47d8-b724-b369bac0bec7 req-efb0dd87-8116-423f-8217-de31fa87fd38 service nova] Lock "929cf4cd-6f3c-4988-8f80-d889d3070e1e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.746647] env[63021]: DEBUG oslo_concurrency.lockutils [req-0ae252cb-a68e-47d8-b724-b369bac0bec7 req-efb0dd87-8116-423f-8217-de31fa87fd38 service nova] Lock "929cf4cd-6f3c-4988-8f80-d889d3070e1e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.746808] env[63021]: DEBUG nova.compute.manager [req-0ae252cb-a68e-47d8-b724-b369bac0bec7 req-efb0dd87-8116-423f-8217-de31fa87fd38 service nova] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] No waiting events found dispatching network-vif-plugged-034fe798-4091-44e1-9dcd-618125f3a581 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 804.746961] env[63021]: WARNING nova.compute.manager [req-0ae252cb-a68e-47d8-b724-b369bac0bec7 req-efb0dd87-8116-423f-8217-de31fa87fd38 service nova] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Received unexpected event network-vif-plugged-034fe798-4091-44e1-9dcd-618125f3a581 for instance with vm_state building and task_state spawning. [ 804.758566] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.430s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.759078] env[63021]: DEBUG nova.compute.manager [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 804.764192] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.661s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.764192] env[63021]: INFO nova.compute.claims [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 804.963955] env[63021]: DEBUG nova.network.neutron [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Successfully updated port: 034fe798-4091-44e1-9dcd-618125f3a581 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 805.204691] env[63021]: DEBUG oslo_vmware.api [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293595, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071427} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.204979] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 805.205791] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4794ec27-bd4d-4594-83c9-b3b7c1d9787e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.226216] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Reconfiguring VM instance instance-00000035 to attach disk [datastore1] 9a6720ef-5307-44b7-b566-8d553c9b5384/9a6720ef-5307-44b7-b566-8d553c9b5384.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 805.227044] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4cc0709f-059e-4e4d-ae4b-bce30b6e4cfe {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.254182] env[63021]: DEBUG oslo_vmware.api [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Waiting for the task: (returnval){ [ 805.254182] env[63021]: value = "task-1293596" [ 805.254182] env[63021]: _type = "Task" [ 805.254182] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.266977] env[63021]: DEBUG nova.compute.utils [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 805.269386] env[63021]: DEBUG oslo_vmware.api [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293596, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.270722] env[63021]: DEBUG nova.compute.manager [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 805.270722] env[63021]: DEBUG nova.network.neutron [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 805.335499] env[63021]: DEBUG nova.policy [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '14d6865c04e44c069a31ca014c2a332d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3f3d9b0b38e647a6b13f2a9af471147b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 805.421228] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquiring lock "f0154308-e6a8-4321-a082-99291344664e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.421463] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Lock "f0154308-e6a8-4321-a082-99291344664e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.444386] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquiring lock "44dadf8e-a727-448f-887d-6408f7ebfbc6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.444668] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Lock "44dadf8e-a727-448f-887d-6408f7ebfbc6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.470228] env[63021]: DEBUG oslo_concurrency.lockutils [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Acquiring lock "refresh_cache-929cf4cd-6f3c-4988-8f80-d889d3070e1e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.470228] env[63021]: DEBUG oslo_concurrency.lockutils [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Acquired lock "refresh_cache-929cf4cd-6f3c-4988-8f80-d889d3070e1e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.470228] env[63021]: DEBUG nova.network.neutron [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 805.477699] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquiring lock "50a1269e-4d83-4cf7-ba14-3b1afb3c43ed" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.478320] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Lock "50a1269e-4d83-4cf7-ba14-3b1afb3c43ed" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.674060] env[63021]: DEBUG nova.network.neutron [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Successfully created port: 9ba9d35e-148e-45b3-a105-3ed457c3803c {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 805.772336] env[63021]: DEBUG nova.compute.manager [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 805.779036] env[63021]: DEBUG oslo_vmware.api [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293596, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.030746] env[63021]: DEBUG nova.network.neutron [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 806.181081] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14d717b6-2523-4285-9478-ecef09a9458d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.189625] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e697990-4e22-4d7d-a9e8-e8b3108526da {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.222233] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01e8b77b-9aa1-4b52-988b-81ba4d7a13db {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.231029] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e246fbc-aac0-4dae-822d-5f4145bded43 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.249069] env[63021]: DEBUG nova.compute.provider_tree [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 806.264597] env[63021]: DEBUG oslo_vmware.api [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293596, 'name': ReconfigVM_Task, 'duration_secs': 0.593897} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.264865] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Reconfigured VM instance instance-00000035 to attach disk [datastore1] 9a6720ef-5307-44b7-b566-8d553c9b5384/9a6720ef-5307-44b7-b566-8d553c9b5384.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 806.265470] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-efc405ea-24f0-41ac-bbbc-a77086fedc2b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.272077] env[63021]: DEBUG oslo_vmware.api [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Waiting for the task: (returnval){ [ 806.272077] env[63021]: value = "task-1293597" [ 806.272077] env[63021]: _type = "Task" [ 806.272077] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.286539] env[63021]: DEBUG oslo_vmware.api [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293597, 'name': Rename_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.313471] env[63021]: DEBUG nova.network.neutron [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Updating instance_info_cache with network_info: [{"id": "034fe798-4091-44e1-9dcd-618125f3a581", "address": "fa:16:3e:c9:81:46", "network": {"id": "5a98d1ac-1821-427f-ab85-bbe4ad812007", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-548917913-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f1a621cb616420bb530cc905d86948b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "674802e7-b847-4bef-a7a8-f90ac7a3a0a7", "external-id": "nsx-vlan-transportzone-953", "segmentation_id": 953, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap034fe798-40", "ovs_interfaceid": "034fe798-4091-44e1-9dcd-618125f3a581", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.752691] env[63021]: DEBUG nova.scheduler.client.report [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 806.785278] env[63021]: DEBUG nova.compute.manager [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 806.787412] env[63021]: DEBUG oslo_vmware.api [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293597, 'name': Rename_Task, 'duration_secs': 0.127218} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.787834] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 806.788079] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1893d72f-a73e-48b2-9731-2ba2768eb8eb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.794342] env[63021]: DEBUG oslo_vmware.api [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Waiting for the task: (returnval){ [ 806.794342] env[63021]: value = "task-1293598" [ 806.794342] env[63021]: _type = "Task" [ 806.794342] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.801701] env[63021]: DEBUG oslo_vmware.api [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293598, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.809942] env[63021]: DEBUG nova.virt.hardware [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 806.810175] env[63021]: DEBUG nova.virt.hardware [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 806.810338] env[63021]: DEBUG nova.virt.hardware [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 806.810493] env[63021]: DEBUG nova.virt.hardware [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 806.810633] env[63021]: DEBUG nova.virt.hardware [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 806.810777] env[63021]: DEBUG nova.virt.hardware [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 806.810979] env[63021]: DEBUG nova.virt.hardware [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 806.811187] env[63021]: DEBUG nova.virt.hardware [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 806.811306] env[63021]: DEBUG nova.virt.hardware [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 806.811507] env[63021]: DEBUG nova.virt.hardware [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 806.811695] env[63021]: DEBUG nova.virt.hardware [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 806.812449] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-145445c3-21c9-4657-9d16-de27dc54d5db {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.816343] env[63021]: DEBUG oslo_concurrency.lockutils [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Releasing lock "refresh_cache-929cf4cd-6f3c-4988-8f80-d889d3070e1e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.816724] env[63021]: DEBUG nova.compute.manager [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Instance network_info: |[{"id": "034fe798-4091-44e1-9dcd-618125f3a581", "address": "fa:16:3e:c9:81:46", "network": {"id": "5a98d1ac-1821-427f-ab85-bbe4ad812007", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-548917913-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f1a621cb616420bb530cc905d86948b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "674802e7-b847-4bef-a7a8-f90ac7a3a0a7", "external-id": "nsx-vlan-transportzone-953", "segmentation_id": 953, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap034fe798-40", "ovs_interfaceid": "034fe798-4091-44e1-9dcd-618125f3a581", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 806.819044] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c9:81:46', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '674802e7-b847-4bef-a7a8-f90ac7a3a0a7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '034fe798-4091-44e1-9dcd-618125f3a581', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 806.827059] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Creating folder: Project (5f1a621cb616420bb530cc905d86948b). Parent ref: group-v277447. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 806.827371] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-34053190-2c2c-407a-8fc8-a9de9144b238 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.830242] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d7b97d1-d706-4f41-8e88-a1206b06ed79 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.845705] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Created folder: Project (5f1a621cb616420bb530cc905d86948b) in parent group-v277447. [ 806.845899] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Creating folder: Instances. Parent ref: group-v277466. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 806.846146] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8dbfc32a-5838-4bb3-b43b-6facac70bb80 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.855611] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Created folder: Instances in parent group-v277466. [ 806.855830] env[63021]: DEBUG oslo.service.loopingcall [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 806.856031] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 806.856241] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-05929ca5-3b99-45f9-aee1-b36cfa7dd769 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.872010] env[63021]: DEBUG nova.compute.manager [req-bb6b953a-54ed-4cbd-9c5d-4eddac317e46 req-5d5192d7-503b-4251-ab25-0f1fd015a08d service nova] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Received event network-changed-034fe798-4091-44e1-9dcd-618125f3a581 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 806.872248] env[63021]: DEBUG nova.compute.manager [req-bb6b953a-54ed-4cbd-9c5d-4eddac317e46 req-5d5192d7-503b-4251-ab25-0f1fd015a08d service nova] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Refreshing instance network info cache due to event network-changed-034fe798-4091-44e1-9dcd-618125f3a581. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 806.872465] env[63021]: DEBUG oslo_concurrency.lockutils [req-bb6b953a-54ed-4cbd-9c5d-4eddac317e46 req-5d5192d7-503b-4251-ab25-0f1fd015a08d service nova] Acquiring lock "refresh_cache-929cf4cd-6f3c-4988-8f80-d889d3070e1e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.872604] env[63021]: DEBUG oslo_concurrency.lockutils [req-bb6b953a-54ed-4cbd-9c5d-4eddac317e46 req-5d5192d7-503b-4251-ab25-0f1fd015a08d service nova] Acquired lock "refresh_cache-929cf4cd-6f3c-4988-8f80-d889d3070e1e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.872759] env[63021]: DEBUG nova.network.neutron [req-bb6b953a-54ed-4cbd-9c5d-4eddac317e46 req-5d5192d7-503b-4251-ab25-0f1fd015a08d service nova] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Refreshing network info cache for port 034fe798-4091-44e1-9dcd-618125f3a581 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 806.877932] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 806.877932] env[63021]: value = "task-1293601" [ 806.877932] env[63021]: _type = "Task" [ 806.877932] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.886195] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293601, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.205916] env[63021]: DEBUG nova.network.neutron [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Successfully updated port: 9ba9d35e-148e-45b3-a105-3ed457c3803c {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 807.257342] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.495s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.257983] env[63021]: DEBUG nova.compute.manager [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 807.264090] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.398s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.265732] env[63021]: INFO nova.compute.claims [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 808.041391] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Acquiring lock "refresh_cache-7c6d7839-616d-41f5-a909-f8c626477a5e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.041541] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Acquired lock "refresh_cache-7c6d7839-616d-41f5-a909-f8c626477a5e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.041686] env[63021]: DEBUG nova.network.neutron [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 808.043595] env[63021]: DEBUG nova.compute.utils [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 808.049973] env[63021]: DEBUG nova.compute.manager [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 808.050157] env[63021]: DEBUG nova.network.neutron [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 808.056062] env[63021]: DEBUG oslo_vmware.api [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293598, 'name': PowerOnVM_Task, 'duration_secs': 0.548576} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.056965] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 808.057180] env[63021]: INFO nova.compute.manager [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Took 6.07 seconds to spawn the instance on the hypervisor. [ 808.057350] env[63021]: DEBUG nova.compute.manager [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 808.058323] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c60b3e6-dde9-4b03-8663-a3930a4834e9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.064556] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293601, 'name': CreateVM_Task, 'duration_secs': 0.535631} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.064655] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 808.081694] env[63021]: DEBUG oslo_concurrency.lockutils [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.081885] env[63021]: DEBUG oslo_concurrency.lockutils [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.082539] env[63021]: DEBUG oslo_concurrency.lockutils [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 808.082539] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2664cace-3403-4b3d-8e44-144c3af2648b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.087452] env[63021]: DEBUG oslo_vmware.api [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Waiting for the task: (returnval){ [ 808.087452] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52cf0113-91f4-91ce-1a75-b1bd2826306e" [ 808.087452] env[63021]: _type = "Task" [ 808.087452] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.094846] env[63021]: DEBUG oslo_vmware.api [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52cf0113-91f4-91ce-1a75-b1bd2826306e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.266560] env[63021]: DEBUG nova.policy [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'efef580358fd48b3a9b5222ddfbee6c8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '14ffe281cb444a0d8d2983f7fec61faa', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 808.331239] env[63021]: DEBUG nova.network.neutron [req-bb6b953a-54ed-4cbd-9c5d-4eddac317e46 req-5d5192d7-503b-4251-ab25-0f1fd015a08d service nova] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Updated VIF entry in instance network info cache for port 034fe798-4091-44e1-9dcd-618125f3a581. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 808.331655] env[63021]: DEBUG nova.network.neutron [req-bb6b953a-54ed-4cbd-9c5d-4eddac317e46 req-5d5192d7-503b-4251-ab25-0f1fd015a08d service nova] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Updating instance_info_cache with network_info: [{"id": "034fe798-4091-44e1-9dcd-618125f3a581", "address": "fa:16:3e:c9:81:46", "network": {"id": "5a98d1ac-1821-427f-ab85-bbe4ad812007", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-548917913-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f1a621cb616420bb530cc905d86948b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "674802e7-b847-4bef-a7a8-f90ac7a3a0a7", "external-id": "nsx-vlan-transportzone-953", "segmentation_id": 953, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap034fe798-40", "ovs_interfaceid": "034fe798-4091-44e1-9dcd-618125f3a581", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.557874] env[63021]: DEBUG nova.compute.manager [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 808.584019] env[63021]: INFO nova.compute.manager [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Took 34.19 seconds to build instance. [ 808.593442] env[63021]: DEBUG nova.network.neutron [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 808.604648] env[63021]: DEBUG oslo_vmware.api [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52cf0113-91f4-91ce-1a75-b1bd2826306e, 'name': SearchDatastore_Task, 'duration_secs': 0.009446} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.605160] env[63021]: DEBUG oslo_concurrency.lockutils [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.605396] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 808.605626] env[63021]: DEBUG oslo_concurrency.lockutils [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.605773] env[63021]: DEBUG oslo_concurrency.lockutils [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.605964] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 808.606501] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1ff6fa0c-f7a3-44d6-9b1b-bfa1b085b435 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.617997] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 808.619150] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 808.620095] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d797acd-0a32-4aa3-b3c1-9c5b5ba9c56d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.627105] env[63021]: DEBUG oslo_vmware.api [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Waiting for the task: (returnval){ [ 808.627105] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]528bfa09-3cef-10e3-ec03-8d285a15d34e" [ 808.627105] env[63021]: _type = "Task" [ 808.627105] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.633656] env[63021]: DEBUG oslo_vmware.api [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]528bfa09-3cef-10e3-ec03-8d285a15d34e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.675299] env[63021]: DEBUG nova.network.neutron [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Successfully created port: 7b37c85f-8b13-40cf-9407-1c618513578c {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 808.765604] env[63021]: DEBUG nova.network.neutron [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Updating instance_info_cache with network_info: [{"id": "9ba9d35e-148e-45b3-a105-3ed457c3803c", "address": "fa:16:3e:6a:bd:ae", "network": {"id": "d67a9f2c-02da-411f-81dd-e2d702d6dd8b", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.35", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "cc50c08b679248798f76b07c1530ed7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "503991c4-44d0-42d9-aa03-5259331f1051", "external-id": "nsx-vlan-transportzone-3", "segmentation_id": 3, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9ba9d35e-14", "ovs_interfaceid": "9ba9d35e-148e-45b3-a105-3ed457c3803c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.835038] env[63021]: DEBUG oslo_concurrency.lockutils [req-bb6b953a-54ed-4cbd-9c5d-4eddac317e46 req-5d5192d7-503b-4251-ab25-0f1fd015a08d service nova] Releasing lock "refresh_cache-929cf4cd-6f3c-4988-8f80-d889d3070e1e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.897574] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad9adb9d-aed9-425a-ac5b-3e7fbd66a070 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.905451] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a24341e4-26dc-49af-b29e-8b1a0b5edc7c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.934556] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-549d6afc-fd4e-4d06-9939-fb9cf513949a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.942081] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f7349db-721b-4060-9143-7f061883740e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.955852] env[63021]: DEBUG nova.compute.provider_tree [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 808.964908] env[63021]: DEBUG nova.compute.manager [req-0dfee5ec-7d57-432a-bacc-cc6ad77184eb req-556c230f-1e05-4f51-9030-665d73cb04d3 service nova] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Received event network-vif-plugged-9ba9d35e-148e-45b3-a105-3ed457c3803c {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 808.965152] env[63021]: DEBUG oslo_concurrency.lockutils [req-0dfee5ec-7d57-432a-bacc-cc6ad77184eb req-556c230f-1e05-4f51-9030-665d73cb04d3 service nova] Acquiring lock "7c6d7839-616d-41f5-a909-f8c626477a5e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.965365] env[63021]: DEBUG oslo_concurrency.lockutils [req-0dfee5ec-7d57-432a-bacc-cc6ad77184eb req-556c230f-1e05-4f51-9030-665d73cb04d3 service nova] Lock "7c6d7839-616d-41f5-a909-f8c626477a5e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.965541] env[63021]: DEBUG oslo_concurrency.lockutils [req-0dfee5ec-7d57-432a-bacc-cc6ad77184eb req-556c230f-1e05-4f51-9030-665d73cb04d3 service nova] Lock "7c6d7839-616d-41f5-a909-f8c626477a5e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.965698] env[63021]: DEBUG nova.compute.manager [req-0dfee5ec-7d57-432a-bacc-cc6ad77184eb req-556c230f-1e05-4f51-9030-665d73cb04d3 service nova] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] No waiting events found dispatching network-vif-plugged-9ba9d35e-148e-45b3-a105-3ed457c3803c {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 808.965864] env[63021]: WARNING nova.compute.manager [req-0dfee5ec-7d57-432a-bacc-cc6ad77184eb req-556c230f-1e05-4f51-9030-665d73cb04d3 service nova] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Received unexpected event network-vif-plugged-9ba9d35e-148e-45b3-a105-3ed457c3803c for instance with vm_state building and task_state spawning. [ 808.966016] env[63021]: DEBUG nova.compute.manager [req-0dfee5ec-7d57-432a-bacc-cc6ad77184eb req-556c230f-1e05-4f51-9030-665d73cb04d3 service nova] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Received event network-changed-9ba9d35e-148e-45b3-a105-3ed457c3803c {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 808.966451] env[63021]: DEBUG nova.compute.manager [req-0dfee5ec-7d57-432a-bacc-cc6ad77184eb req-556c230f-1e05-4f51-9030-665d73cb04d3 service nova] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Refreshing instance network info cache due to event network-changed-9ba9d35e-148e-45b3-a105-3ed457c3803c. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 808.966768] env[63021]: DEBUG oslo_concurrency.lockutils [req-0dfee5ec-7d57-432a-bacc-cc6ad77184eb req-556c230f-1e05-4f51-9030-665d73cb04d3 service nova] Acquiring lock "refresh_cache-7c6d7839-616d-41f5-a909-f8c626477a5e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.083158] env[63021]: DEBUG oslo_concurrency.lockutils [None req-227027e2-d081-4908-999f-b743373450da tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Lock "9a6720ef-5307-44b7-b566-8d553c9b5384" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 111.550s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.136181] env[63021]: DEBUG oslo_vmware.api [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]528bfa09-3cef-10e3-ec03-8d285a15d34e, 'name': SearchDatastore_Task, 'duration_secs': 0.009402} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.137010] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-712d5089-b836-44b4-a024-128c6a69302d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.141939] env[63021]: DEBUG oslo_vmware.api [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Waiting for the task: (returnval){ [ 809.141939] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5202d5da-cfe0-36db-b70b-06e9cc49cab7" [ 809.141939] env[63021]: _type = "Task" [ 809.141939] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.150485] env[63021]: DEBUG oslo_vmware.api [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5202d5da-cfe0-36db-b70b-06e9cc49cab7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.267779] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Releasing lock "refresh_cache-7c6d7839-616d-41f5-a909-f8c626477a5e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.268202] env[63021]: DEBUG nova.compute.manager [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Instance network_info: |[{"id": "9ba9d35e-148e-45b3-a105-3ed457c3803c", "address": "fa:16:3e:6a:bd:ae", "network": {"id": "d67a9f2c-02da-411f-81dd-e2d702d6dd8b", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.35", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "cc50c08b679248798f76b07c1530ed7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "503991c4-44d0-42d9-aa03-5259331f1051", "external-id": "nsx-vlan-transportzone-3", "segmentation_id": 3, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9ba9d35e-14", "ovs_interfaceid": "9ba9d35e-148e-45b3-a105-3ed457c3803c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 809.268544] env[63021]: DEBUG oslo_concurrency.lockutils [req-0dfee5ec-7d57-432a-bacc-cc6ad77184eb req-556c230f-1e05-4f51-9030-665d73cb04d3 service nova] Acquired lock "refresh_cache-7c6d7839-616d-41f5-a909-f8c626477a5e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.268731] env[63021]: DEBUG nova.network.neutron [req-0dfee5ec-7d57-432a-bacc-cc6ad77184eb req-556c230f-1e05-4f51-9030-665d73cb04d3 service nova] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Refreshing network info cache for port 9ba9d35e-148e-45b3-a105-3ed457c3803c {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 809.270041] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6a:bd:ae', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '503991c4-44d0-42d9-aa03-5259331f1051', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9ba9d35e-148e-45b3-a105-3ed457c3803c', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 809.281622] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Creating folder: Project (3f3d9b0b38e647a6b13f2a9af471147b). Parent ref: group-v277447. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 809.285959] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bba04be5-f300-4b3b-9428-59a4940e1da0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.297888] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Created folder: Project (3f3d9b0b38e647a6b13f2a9af471147b) in parent group-v277447. [ 809.298085] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Creating folder: Instances. Parent ref: group-v277469. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 809.298332] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a7551352-9b91-49cc-948d-9a72f55a7ee9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.307239] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Created folder: Instances in parent group-v277469. [ 809.307485] env[63021]: DEBUG oslo.service.loopingcall [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 809.308328] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 809.308328] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-74ad0fb0-7f82-4c6b-935b-74e65b4fc987 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.325032] env[63021]: INFO nova.compute.manager [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Rebuilding instance [ 809.331026] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 809.331026] env[63021]: value = "task-1293604" [ 809.331026] env[63021]: _type = "Task" [ 809.331026] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.339997] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293604, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.373842] env[63021]: DEBUG nova.compute.manager [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 809.374730] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da2b49eb-b3cb-4e3c-b2da-c45a6149e100 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.458521] env[63021]: DEBUG nova.scheduler.client.report [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 809.500687] env[63021]: DEBUG nova.network.neutron [req-0dfee5ec-7d57-432a-bacc-cc6ad77184eb req-556c230f-1e05-4f51-9030-665d73cb04d3 service nova] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Updated VIF entry in instance network info cache for port 9ba9d35e-148e-45b3-a105-3ed457c3803c. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 809.501063] env[63021]: DEBUG nova.network.neutron [req-0dfee5ec-7d57-432a-bacc-cc6ad77184eb req-556c230f-1e05-4f51-9030-665d73cb04d3 service nova] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Updating instance_info_cache with network_info: [{"id": "9ba9d35e-148e-45b3-a105-3ed457c3803c", "address": "fa:16:3e:6a:bd:ae", "network": {"id": "d67a9f2c-02da-411f-81dd-e2d702d6dd8b", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.35", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "cc50c08b679248798f76b07c1530ed7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "503991c4-44d0-42d9-aa03-5259331f1051", "external-id": "nsx-vlan-transportzone-3", "segmentation_id": 3, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9ba9d35e-14", "ovs_interfaceid": "9ba9d35e-148e-45b3-a105-3ed457c3803c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.572182] env[63021]: DEBUG nova.compute.manager [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 809.588688] env[63021]: DEBUG nova.compute.manager [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 809.594372] env[63021]: DEBUG nova.virt.hardware [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 809.594372] env[63021]: DEBUG nova.virt.hardware [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 809.594372] env[63021]: DEBUG nova.virt.hardware [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 809.594509] env[63021]: DEBUG nova.virt.hardware [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 809.594509] env[63021]: DEBUG nova.virt.hardware [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 809.594509] env[63021]: DEBUG nova.virt.hardware [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 809.594723] env[63021]: DEBUG nova.virt.hardware [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 809.594723] env[63021]: DEBUG nova.virt.hardware [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 809.594867] env[63021]: DEBUG nova.virt.hardware [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 809.595043] env[63021]: DEBUG nova.virt.hardware [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 809.595312] env[63021]: DEBUG nova.virt.hardware [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 809.596115] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6e51061-507a-4392-8e73-209a2fe283e1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.604718] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f9cf7bc-da8a-4e86-bf5d-b186d8177424 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.651217] env[63021]: DEBUG oslo_vmware.api [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5202d5da-cfe0-36db-b70b-06e9cc49cab7, 'name': SearchDatastore_Task, 'duration_secs': 0.010728} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.651494] env[63021]: DEBUG oslo_concurrency.lockutils [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.651720] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 929cf4cd-6f3c-4988-8f80-d889d3070e1e/929cf4cd-6f3c-4988-8f80-d889d3070e1e.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 809.652231] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a1d0e89b-618b-4e10-b03a-b1d941f4089b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.658153] env[63021]: DEBUG oslo_vmware.api [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Waiting for the task: (returnval){ [ 809.658153] env[63021]: value = "task-1293605" [ 809.658153] env[63021]: _type = "Task" [ 809.658153] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.666019] env[63021]: DEBUG oslo_vmware.api [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Task: {'id': task-1293605, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.840507] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293604, 'name': CreateVM_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.886581] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 809.886908] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c16ed81e-91e8-44b3-96c9-1c55847733e4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.894497] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Waiting for the task: (returnval){ [ 809.894497] env[63021]: value = "task-1293606" [ 809.894497] env[63021]: _type = "Task" [ 809.894497] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.905072] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293606, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.964394] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.700s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.965088] env[63021]: DEBUG nova.compute.manager [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 809.968191] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 15.836s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.968393] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.968568] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63021) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 809.968886] env[63021]: DEBUG oslo_concurrency.lockutils [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.744s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.970519] env[63021]: INFO nova.compute.claims [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 809.974047] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00c0f5a8-797b-4b36-a0e3-0216f4dfc4c7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.983310] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0763450-1288-4a29-b428-ab1f9da10b98 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.000418] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5edcf57-2006-492f-83c3-070a44c68bdb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.003437] env[63021]: DEBUG oslo_concurrency.lockutils [req-0dfee5ec-7d57-432a-bacc-cc6ad77184eb req-556c230f-1e05-4f51-9030-665d73cb04d3 service nova] Releasing lock "refresh_cache-7c6d7839-616d-41f5-a909-f8c626477a5e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.008819] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb77c9f4-500c-4df8-87a7-471a1e03a3a2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.039170] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181529MB free_disk=139GB free_vcpus=48 pci_devices=None {{(pid=63021) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 810.039340] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.105652] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.168344] env[63021]: DEBUG oslo_vmware.api [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Task: {'id': task-1293605, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.319509] env[63021]: DEBUG nova.compute.manager [req-96d20214-9947-4561-b4aa-1d9ca846921b req-c939d7e3-9550-489e-99ab-e0e709883e6c service nova] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Received event network-vif-plugged-7b37c85f-8b13-40cf-9407-1c618513578c {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 810.319737] env[63021]: DEBUG oslo_concurrency.lockutils [req-96d20214-9947-4561-b4aa-1d9ca846921b req-c939d7e3-9550-489e-99ab-e0e709883e6c service nova] Acquiring lock "bef5c7cb-be2f-4be7-b72c-4a066d473dd0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.319947] env[63021]: DEBUG oslo_concurrency.lockutils [req-96d20214-9947-4561-b4aa-1d9ca846921b req-c939d7e3-9550-489e-99ab-e0e709883e6c service nova] Lock "bef5c7cb-be2f-4be7-b72c-4a066d473dd0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.320124] env[63021]: DEBUG oslo_concurrency.lockutils [req-96d20214-9947-4561-b4aa-1d9ca846921b req-c939d7e3-9550-489e-99ab-e0e709883e6c service nova] Lock "bef5c7cb-be2f-4be7-b72c-4a066d473dd0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.320295] env[63021]: DEBUG nova.compute.manager [req-96d20214-9947-4561-b4aa-1d9ca846921b req-c939d7e3-9550-489e-99ab-e0e709883e6c service nova] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] No waiting events found dispatching network-vif-plugged-7b37c85f-8b13-40cf-9407-1c618513578c {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 810.320455] env[63021]: WARNING nova.compute.manager [req-96d20214-9947-4561-b4aa-1d9ca846921b req-c939d7e3-9550-489e-99ab-e0e709883e6c service nova] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Received unexpected event network-vif-plugged-7b37c85f-8b13-40cf-9407-1c618513578c for instance with vm_state building and task_state spawning. [ 810.342639] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293604, 'name': CreateVM_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.406543] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293606, 'name': PowerOffVM_Task, 'duration_secs': 0.12777} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.406842] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 810.407069] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 810.407819] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05d588ed-9282-4a32-b3de-3dc0467d4960 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.416920] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 810.416990] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-afe2f9c5-a181-4ba3-b7f4-06ebbac2a9a8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.447019] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 810.447019] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 810.447019] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Deleting the datastore file [datastore1] 9a6720ef-5307-44b7-b566-8d553c9b5384 {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 810.447019] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0c8e1954-3380-4811-8aea-e7e63c373506 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.453019] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Waiting for the task: (returnval){ [ 810.453019] env[63021]: value = "task-1293608" [ 810.453019] env[63021]: _type = "Task" [ 810.453019] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.464473] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293608, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.476105] env[63021]: DEBUG nova.network.neutron [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Successfully updated port: 7b37c85f-8b13-40cf-9407-1c618513578c {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 810.476105] env[63021]: DEBUG nova.compute.utils [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 810.481546] env[63021]: DEBUG nova.compute.manager [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 810.481729] env[63021]: DEBUG nova.network.neutron [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 810.532932] env[63021]: DEBUG nova.policy [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6ea19453dbdb4251a93dde2576aeff64', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '58693d1e991249f28914a39761cc2627', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 810.670017] env[63021]: DEBUG oslo_vmware.api [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Task: {'id': task-1293605, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.841856] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293604, 'name': CreateVM_Task, 'duration_secs': 1.299402} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.842404] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 810.843096] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.843268] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.843617] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 810.843891] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-18cb4e4c-6b3d-4876-8f47-f109ef8b2c01 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.848919] env[63021]: DEBUG oslo_vmware.api [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Waiting for the task: (returnval){ [ 810.848919] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]523017ce-6442-54fe-0bfb-c3de0933fde7" [ 810.848919] env[63021]: _type = "Task" [ 810.848919] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.857336] env[63021]: DEBUG oslo_vmware.api [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]523017ce-6442-54fe-0bfb-c3de0933fde7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.939844] env[63021]: DEBUG nova.network.neutron [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Successfully created port: 37dcb3f1-2931-40c6-96ba-b31b8fa1a72b {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 810.960809] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293608, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.08525} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.961090] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 810.961277] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 810.961460] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 810.984407] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Acquiring lock "refresh_cache-bef5c7cb-be2f-4be7-b72c-4a066d473dd0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.984407] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Acquired lock "refresh_cache-bef5c7cb-be2f-4be7-b72c-4a066d473dd0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.984407] env[63021]: DEBUG nova.network.neutron [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 810.987983] env[63021]: DEBUG nova.compute.manager [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 811.170446] env[63021]: DEBUG oslo_vmware.api [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Task: {'id': task-1293605, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.473563} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.170751] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 929cf4cd-6f3c-4988-8f80-d889d3070e1e/929cf4cd-6f3c-4988-8f80-d889d3070e1e.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 811.170955] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 811.171210] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8dc38185-39ac-483c-84de-283a7e27a9af {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.177292] env[63021]: DEBUG oslo_vmware.api [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Waiting for the task: (returnval){ [ 811.177292] env[63021]: value = "task-1293609" [ 811.177292] env[63021]: _type = "Task" [ 811.177292] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.187598] env[63021]: DEBUG oslo_vmware.api [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Task: {'id': task-1293609, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.303379] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d5a9eff-fe3b-48fa-ad5a-1d34533a586f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.311363] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a59c6e1-9837-49fe-9ac4-e89319987d43 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.342541] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bae4f0b0-4b10-4a09-ae79-ffa5e0eb8934 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.350025] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07a1cd9e-0526-4644-b6f3-2135a2bec662 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.365455] env[63021]: DEBUG nova.compute.provider_tree [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 811.369913] env[63021]: DEBUG oslo_vmware.api [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]523017ce-6442-54fe-0bfb-c3de0933fde7, 'name': SearchDatastore_Task, 'duration_secs': 0.120538} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.370354] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.370611] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 811.370841] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.370979] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.371162] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 811.371396] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0a46cce4-82d6-4ef6-af71-6aca69824659 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.379836] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 811.380041] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 811.380709] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-616c208b-4615-40ea-9003-1d2004590296 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.386181] env[63021]: DEBUG oslo_vmware.api [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Waiting for the task: (returnval){ [ 811.386181] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52a51aa1-93d2-f3dd-a4d9-6ba6607b7f9f" [ 811.386181] env[63021]: _type = "Task" [ 811.386181] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.393540] env[63021]: DEBUG oslo_vmware.api [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52a51aa1-93d2-f3dd-a4d9-6ba6607b7f9f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.513682] env[63021]: DEBUG nova.network.neutron [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 811.643466] env[63021]: DEBUG nova.network.neutron [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Updating instance_info_cache with network_info: [{"id": "7b37c85f-8b13-40cf-9407-1c618513578c", "address": "fa:16:3e:82:3b:69", "network": {"id": "d030cf3c-4945-49cb-9987-b1844913892e", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1561367713-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "14ffe281cb444a0d8d2983f7fec61faa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "86a35d07-53d3-46b3-92cb-ae34236c0f41", "external-id": "nsx-vlan-transportzone-811", "segmentation_id": 811, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b37c85f-8b", "ovs_interfaceid": "7b37c85f-8b13-40cf-9407-1c618513578c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.687979] env[63021]: DEBUG oslo_vmware.api [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Task: {'id': task-1293609, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074424} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.688281] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 811.690819] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0730b0a-eba7-4910-b05f-c71a91aaeac5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.711506] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] 929cf4cd-6f3c-4988-8f80-d889d3070e1e/929cf4cd-6f3c-4988-8f80-d889d3070e1e.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 811.711782] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2c5c8a1-6af6-4af7-88c1-e18a8821dd8d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.731340] env[63021]: DEBUG oslo_vmware.api [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Waiting for the task: (returnval){ [ 811.731340] env[63021]: value = "task-1293610" [ 811.731340] env[63021]: _type = "Task" [ 811.731340] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.738982] env[63021]: DEBUG oslo_vmware.api [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Task: {'id': task-1293610, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.872144] env[63021]: DEBUG nova.scheduler.client.report [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 811.899118] env[63021]: DEBUG oslo_vmware.api [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52a51aa1-93d2-f3dd-a4d9-6ba6607b7f9f, 'name': SearchDatastore_Task, 'duration_secs': 0.007858} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.899885] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-616fe2bd-e296-4cd7-8f90-66b02d85c927 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.905215] env[63021]: DEBUG oslo_vmware.api [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Waiting for the task: (returnval){ [ 811.905215] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52c69679-b06a-308a-a8c7-fea288668847" [ 811.905215] env[63021]: _type = "Task" [ 811.905215] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.912697] env[63021]: DEBUG oslo_vmware.api [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52c69679-b06a-308a-a8c7-fea288668847, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.997695] env[63021]: DEBUG nova.virt.hardware [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 811.998016] env[63021]: DEBUG nova.virt.hardware [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 811.998183] env[63021]: DEBUG nova.virt.hardware [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 811.998367] env[63021]: DEBUG nova.virt.hardware [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 811.998509] env[63021]: DEBUG nova.virt.hardware [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 811.998652] env[63021]: DEBUG nova.virt.hardware [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 811.998859] env[63021]: DEBUG nova.virt.hardware [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 811.999017] env[63021]: DEBUG nova.virt.hardware [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 811.999184] env[63021]: DEBUG nova.virt.hardware [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 811.999580] env[63021]: DEBUG nova.virt.hardware [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 811.999764] env[63021]: DEBUG nova.virt.hardware [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 812.000648] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4f1b7da-0e01-4639-88ac-c696208bef45 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.004039] env[63021]: DEBUG nova.compute.manager [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 812.011548] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7643f71-34c8-4aa8-87ce-3a8181c285ab {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.025925] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Instance VIF info [] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 812.031363] env[63021]: DEBUG oslo.service.loopingcall [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 812.033465] env[63021]: DEBUG nova.virt.hardware [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 812.033710] env[63021]: DEBUG nova.virt.hardware [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 812.033868] env[63021]: DEBUG nova.virt.hardware [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 812.034059] env[63021]: DEBUG nova.virt.hardware [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 812.034204] env[63021]: DEBUG nova.virt.hardware [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 812.034347] env[63021]: DEBUG nova.virt.hardware [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 812.034547] env[63021]: DEBUG nova.virt.hardware [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 812.034701] env[63021]: DEBUG nova.virt.hardware [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 812.034859] env[63021]: DEBUG nova.virt.hardware [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 812.035022] env[63021]: DEBUG nova.virt.hardware [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 812.035192] env[63021]: DEBUG nova.virt.hardware [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 812.035746] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 812.036450] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d72b3c6a-1e18-4ed4-96ab-4536c28bc8cb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.038866] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a4f3dfd7-a8a0-48d7-8930-19f7aead3a54 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.057839] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40443687-ca9f-4b9f-9fb7-7fc5c8392402 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.061058] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 812.061058] env[63021]: value = "task-1293611" [ 812.061058] env[63021]: _type = "Task" [ 812.061058] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.076392] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293611, 'name': CreateVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.146329] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Releasing lock "refresh_cache-bef5c7cb-be2f-4be7-b72c-4a066d473dd0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.146677] env[63021]: DEBUG nova.compute.manager [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Instance network_info: |[{"id": "7b37c85f-8b13-40cf-9407-1c618513578c", "address": "fa:16:3e:82:3b:69", "network": {"id": "d030cf3c-4945-49cb-9987-b1844913892e", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1561367713-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "14ffe281cb444a0d8d2983f7fec61faa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "86a35d07-53d3-46b3-92cb-ae34236c0f41", "external-id": "nsx-vlan-transportzone-811", "segmentation_id": 811, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b37c85f-8b", "ovs_interfaceid": "7b37c85f-8b13-40cf-9407-1c618513578c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 812.147114] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:82:3b:69', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '86a35d07-53d3-46b3-92cb-ae34236c0f41', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7b37c85f-8b13-40cf-9407-1c618513578c', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 812.155145] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Creating folder: Project (14ffe281cb444a0d8d2983f7fec61faa). Parent ref: group-v277447. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 812.155444] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3f48677e-d884-4951-a7a7-4c1708c2f0d4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.167759] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Created folder: Project (14ffe281cb444a0d8d2983f7fec61faa) in parent group-v277447. [ 812.167982] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Creating folder: Instances. Parent ref: group-v277473. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 812.168240] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-13d97abe-95d8-4c0d-999b-a7e3613213e5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.176975] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Created folder: Instances in parent group-v277473. [ 812.177217] env[63021]: DEBUG oslo.service.loopingcall [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 812.177436] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 812.177656] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f0dbf8e3-0605-44c8-bebb-77e22586a66a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.196885] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 812.196885] env[63021]: value = "task-1293614" [ 812.196885] env[63021]: _type = "Task" [ 812.196885] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.204454] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293614, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.241100] env[63021]: DEBUG oslo_vmware.api [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Task: {'id': task-1293610, 'name': ReconfigVM_Task, 'duration_secs': 0.290769} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.241463] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Reconfigured VM instance instance-00000036 to attach disk [datastore1] 929cf4cd-6f3c-4988-8f80-d889d3070e1e/929cf4cd-6f3c-4988-8f80-d889d3070e1e.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 812.242215] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-77d49056-a5c5-4b1f-bc6e-19ac3b29cd96 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.249583] env[63021]: DEBUG oslo_vmware.api [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Waiting for the task: (returnval){ [ 812.249583] env[63021]: value = "task-1293615" [ 812.249583] env[63021]: _type = "Task" [ 812.249583] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.259920] env[63021]: DEBUG oslo_vmware.api [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Task: {'id': task-1293615, 'name': Rename_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.377160] env[63021]: DEBUG oslo_concurrency.lockutils [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.408s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.378124] env[63021]: DEBUG nova.compute.manager [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 812.381379] env[63021]: DEBUG oslo_concurrency.lockutils [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.707s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.416638] env[63021]: DEBUG oslo_vmware.api [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52c69679-b06a-308a-a8c7-fea288668847, 'name': SearchDatastore_Task, 'duration_secs': 0.008044} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.417662] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.417916] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 7c6d7839-616d-41f5-a909-f8c626477a5e/7c6d7839-616d-41f5-a909-f8c626477a5e.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 812.418665] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ac6a2d24-f9f0-4b72-bbbe-5e67e044bf15 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.425440] env[63021]: DEBUG oslo_vmware.api [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Waiting for the task: (returnval){ [ 812.425440] env[63021]: value = "task-1293616" [ 812.425440] env[63021]: _type = "Task" [ 812.425440] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.436920] env[63021]: DEBUG oslo_vmware.api [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293616, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.492415] env[63021]: DEBUG nova.compute.manager [req-a7a6ccf3-9399-428a-be4d-6af3024c556b req-5b52c103-abf9-447f-b1fd-1abdd4f3838c service nova] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Received event network-changed-7b37c85f-8b13-40cf-9407-1c618513578c {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 812.492605] env[63021]: DEBUG nova.compute.manager [req-a7a6ccf3-9399-428a-be4d-6af3024c556b req-5b52c103-abf9-447f-b1fd-1abdd4f3838c service nova] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Refreshing instance network info cache due to event network-changed-7b37c85f-8b13-40cf-9407-1c618513578c. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 812.492817] env[63021]: DEBUG oslo_concurrency.lockutils [req-a7a6ccf3-9399-428a-be4d-6af3024c556b req-5b52c103-abf9-447f-b1fd-1abdd4f3838c service nova] Acquiring lock "refresh_cache-bef5c7cb-be2f-4be7-b72c-4a066d473dd0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.493128] env[63021]: DEBUG oslo_concurrency.lockutils [req-a7a6ccf3-9399-428a-be4d-6af3024c556b req-5b52c103-abf9-447f-b1fd-1abdd4f3838c service nova] Acquired lock "refresh_cache-bef5c7cb-be2f-4be7-b72c-4a066d473dd0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.493128] env[63021]: DEBUG nova.network.neutron [req-a7a6ccf3-9399-428a-be4d-6af3024c556b req-5b52c103-abf9-447f-b1fd-1abdd4f3838c service nova] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Refreshing network info cache for port 7b37c85f-8b13-40cf-9407-1c618513578c {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 812.575190] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293611, 'name': CreateVM_Task, 'duration_secs': 0.321805} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.575279] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 812.576026] env[63021]: DEBUG oslo_concurrency.lockutils [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.576026] env[63021]: DEBUG oslo_concurrency.lockutils [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.576164] env[63021]: DEBUG oslo_concurrency.lockutils [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 812.577050] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-722efb38-adca-4b2b-9fe7-03f23370dd56 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.582147] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Waiting for the task: (returnval){ [ 812.582147] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52744d9d-f6af-8a17-2746-0e411c6e6b1f" [ 812.582147] env[63021]: _type = "Task" [ 812.582147] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.589286] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52744d9d-f6af-8a17-2746-0e411c6e6b1f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.708307] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293614, 'name': CreateVM_Task, 'duration_secs': 0.375058} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.708471] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 812.709150] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.760367] env[63021]: DEBUG oslo_vmware.api [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Task: {'id': task-1293615, 'name': Rename_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.883157] env[63021]: DEBUG nova.network.neutron [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Successfully updated port: 37dcb3f1-2931-40c6-96ba-b31b8fa1a72b {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 812.885944] env[63021]: DEBUG nova.compute.utils [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 812.890847] env[63021]: DEBUG nova.compute.manager [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 812.891039] env[63021]: DEBUG nova.network.neutron [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 812.936139] env[63021]: DEBUG nova.policy [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e2ce7bd37c324e76bad16423ba0e514a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4a0b1326882b40acb72aad7f6c6e2b0a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 812.942560] env[63021]: DEBUG oslo_vmware.api [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293616, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.099019] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52744d9d-f6af-8a17-2746-0e411c6e6b1f, 'name': SearchDatastore_Task, 'duration_secs': 0.020257} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.099019] env[63021]: DEBUG oslo_concurrency.lockutils [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.099019] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 813.099019] env[63021]: DEBUG oslo_concurrency.lockutils [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.099219] env[63021]: DEBUG oslo_concurrency.lockutils [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.099219] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 813.099219] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.099219] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 813.099219] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eca77e67-559f-4b9e-859b-11be338517f5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.101136] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f55a3a34-c3ef-4959-a121-c6c3f0b286f0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.106160] env[63021]: DEBUG oslo_vmware.api [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Waiting for the task: (returnval){ [ 813.106160] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5267cd3a-a8a5-4ad5-ff62-70dc2614c560" [ 813.106160] env[63021]: _type = "Task" [ 813.106160] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.115364] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 813.115563] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 813.116339] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8835b3a6-dabe-4b04-90ae-0ce78b94f45b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.121856] env[63021]: DEBUG oslo_vmware.api [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5267cd3a-a8a5-4ad5-ff62-70dc2614c560, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.127256] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Waiting for the task: (returnval){ [ 813.127256] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52a6eb2f-1431-ad72-3cfe-63e4722419fd" [ 813.127256] env[63021]: _type = "Task" [ 813.127256] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.135199] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52a6eb2f-1431-ad72-3cfe-63e4722419fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.259958] env[63021]: DEBUG nova.network.neutron [req-a7a6ccf3-9399-428a-be4d-6af3024c556b req-5b52c103-abf9-447f-b1fd-1abdd4f3838c service nova] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Updated VIF entry in instance network info cache for port 7b37c85f-8b13-40cf-9407-1c618513578c. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 813.260281] env[63021]: DEBUG nova.network.neutron [req-a7a6ccf3-9399-428a-be4d-6af3024c556b req-5b52c103-abf9-447f-b1fd-1abdd4f3838c service nova] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Updating instance_info_cache with network_info: [{"id": "7b37c85f-8b13-40cf-9407-1c618513578c", "address": "fa:16:3e:82:3b:69", "network": {"id": "d030cf3c-4945-49cb-9987-b1844913892e", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1561367713-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "14ffe281cb444a0d8d2983f7fec61faa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "86a35d07-53d3-46b3-92cb-ae34236c0f41", "external-id": "nsx-vlan-transportzone-811", "segmentation_id": 811, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b37c85f-8b", "ovs_interfaceid": "7b37c85f-8b13-40cf-9407-1c618513578c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.271272] env[63021]: DEBUG oslo_vmware.api [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Task: {'id': task-1293615, 'name': Rename_Task, 'duration_secs': 0.843767} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.272639] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 813.272879] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ee968b87-188a-45b2-af0c-f7ecb2f74339 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.279826] env[63021]: DEBUG oslo_vmware.api [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Waiting for the task: (returnval){ [ 813.279826] env[63021]: value = "task-1293617" [ 813.279826] env[63021]: _type = "Task" [ 813.279826] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.291868] env[63021]: DEBUG oslo_vmware.api [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Task: {'id': task-1293617, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.328463] env[63021]: DEBUG nova.network.neutron [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Successfully created port: c3780241-4bde-4b91-8bd6-4f42ddc35d07 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 813.342350] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6af01c54-41e9-4782-9d66-538458d646ea {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.350792] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8884f045-e266-4623-8310-8705f39c38a0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.382566] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49d87867-b3e2-49c5-88a3-ce8f82758acb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.385638] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "refresh_cache-36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.385778] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquired lock "refresh_cache-36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.385927] env[63021]: DEBUG nova.network.neutron [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 813.391581] env[63021]: DEBUG nova.compute.manager [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 813.401024] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aa58c32-3043-4755-aff4-c2a4632c98a2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.410636] env[63021]: DEBUG nova.compute.provider_tree [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 813.434692] env[63021]: DEBUG oslo_vmware.api [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293616, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.996878} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.434926] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 7c6d7839-616d-41f5-a909-f8c626477a5e/7c6d7839-616d-41f5-a909-f8c626477a5e.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 813.435151] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 813.435387] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-af557b6f-8a7a-4750-8586-acea44eeada5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.441342] env[63021]: DEBUG oslo_vmware.api [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Waiting for the task: (returnval){ [ 813.441342] env[63021]: value = "task-1293618" [ 813.441342] env[63021]: _type = "Task" [ 813.441342] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.449575] env[63021]: DEBUG oslo_vmware.api [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293618, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.625119] env[63021]: DEBUG oslo_vmware.api [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5267cd3a-a8a5-4ad5-ff62-70dc2614c560, 'name': SearchDatastore_Task, 'duration_secs': 0.077616} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.625459] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.625719] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 813.625960] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.637676] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52a6eb2f-1431-ad72-3cfe-63e4722419fd, 'name': SearchDatastore_Task, 'duration_secs': 0.067103} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.638301] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-966a3a58-e480-4253-a9f9-65c647da5c60 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.643729] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Waiting for the task: (returnval){ [ 813.643729] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52c353f6-e8c1-0de2-523d-ff8d1d4bf639" [ 813.643729] env[63021]: _type = "Task" [ 813.643729] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.652054] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52c353f6-e8c1-0de2-523d-ff8d1d4bf639, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.770930] env[63021]: DEBUG oslo_concurrency.lockutils [req-a7a6ccf3-9399-428a-be4d-6af3024c556b req-5b52c103-abf9-447f-b1fd-1abdd4f3838c service nova] Releasing lock "refresh_cache-bef5c7cb-be2f-4be7-b72c-4a066d473dd0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.790951] env[63021]: DEBUG oslo_vmware.api [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Task: {'id': task-1293617, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.919220] env[63021]: DEBUG nova.scheduler.client.report [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 813.921460] env[63021]: DEBUG nova.network.neutron [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 813.951702] env[63021]: DEBUG oslo_vmware.api [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293618, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063909} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.952042] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 813.953110] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0fd7493-8f27-4c1d-b94a-2d84aea6d442 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.979205] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Reconfiguring VM instance instance-00000037 to attach disk [datastore1] 7c6d7839-616d-41f5-a909-f8c626477a5e/7c6d7839-616d-41f5-a909-f8c626477a5e.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 813.981968] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5d5cd660-bd33-4882-8d8b-9da3d3e61b8b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.004303] env[63021]: DEBUG oslo_vmware.api [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Waiting for the task: (returnval){ [ 814.004303] env[63021]: value = "task-1293619" [ 814.004303] env[63021]: _type = "Task" [ 814.004303] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.012100] env[63021]: DEBUG oslo_vmware.api [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293619, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.110769] env[63021]: DEBUG nova.network.neutron [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Updating instance_info_cache with network_info: [{"id": "37dcb3f1-2931-40c6-96ba-b31b8fa1a72b", "address": "fa:16:3e:57:13:17", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37dcb3f1-29", "ovs_interfaceid": "37dcb3f1-2931-40c6-96ba-b31b8fa1a72b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.156765] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52c353f6-e8c1-0de2-523d-ff8d1d4bf639, 'name': SearchDatastore_Task, 'duration_secs': 0.008561} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.157214] env[63021]: DEBUG oslo_concurrency.lockutils [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.157610] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 9a6720ef-5307-44b7-b566-8d553c9b5384/9a6720ef-5307-44b7-b566-8d553c9b5384.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 814.158214] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.158500] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 814.158820] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1361ae5e-cdb3-48da-acb5-cd5098370b69 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.161345] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-71f46bc0-0624-4db2-91b3-f50660941c4c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.168565] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Waiting for the task: (returnval){ [ 814.168565] env[63021]: value = "task-1293620" [ 814.168565] env[63021]: _type = "Task" [ 814.168565] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.173220] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 814.173220] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 814.173930] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-319be050-2915-4a34-b024-71e03a09f669 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.178845] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293620, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.181604] env[63021]: DEBUG oslo_vmware.api [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Waiting for the task: (returnval){ [ 814.181604] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52411a38-bea5-8277-890c-9c3f70d31326" [ 814.181604] env[63021]: _type = "Task" [ 814.181604] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.189860] env[63021]: DEBUG oslo_vmware.api [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52411a38-bea5-8277-890c-9c3f70d31326, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.289905] env[63021]: DEBUG oslo_vmware.api [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Task: {'id': task-1293617, 'name': PowerOnVM_Task, 'duration_secs': 0.517146} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.290202] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 814.290408] env[63021]: INFO nova.compute.manager [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Took 9.93 seconds to spawn the instance on the hypervisor. [ 814.290585] env[63021]: DEBUG nova.compute.manager [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 814.291335] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea1d0791-c350-4f1b-b794-0cb329c68ee5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.406754] env[63021]: DEBUG nova.compute.manager [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 814.421344] env[63021]: DEBUG oslo_concurrency.lockutils [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.040s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.421971] env[63021]: ERROR nova.compute.manager [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d62d2305-e5cd-4344-a697-444a752049d9, please check neutron logs for more information. [ 814.421971] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Traceback (most recent call last): [ 814.421971] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 814.421971] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] self.driver.spawn(context, instance, image_meta, [ 814.421971] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 814.421971] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 814.421971] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 814.421971] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] vm_ref = self.build_virtual_machine(instance, [ 814.421971] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 814.421971] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] vif_infos = vmwarevif.get_vif_info(self._session, [ 814.421971] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 814.422469] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] for vif in network_info: [ 814.422469] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 814.422469] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] return self._sync_wrapper(fn, *args, **kwargs) [ 814.422469] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 814.422469] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] self.wait() [ 814.422469] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 814.422469] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] self[:] = self._gt.wait() [ 814.422469] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 814.422469] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] return self._exit_event.wait() [ 814.422469] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 814.422469] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] current.throw(*self._exc) [ 814.422469] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 814.422469] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] result = function(*args, **kwargs) [ 814.422802] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 814.422802] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] return func(*args, **kwargs) [ 814.422802] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 814.422802] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] raise e [ 814.422802] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 814.422802] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] nwinfo = self.network_api.allocate_for_instance( [ 814.422802] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 814.422802] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] created_port_ids = self._update_ports_for_instance( [ 814.422802] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 814.422802] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] with excutils.save_and_reraise_exception(): [ 814.422802] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 814.422802] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] self.force_reraise() [ 814.422802] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 814.423095] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] raise self.value [ 814.423095] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 814.423095] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] updated_port = self._update_port( [ 814.423095] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 814.423095] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] _ensure_no_port_binding_failure(port) [ 814.423095] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 814.423095] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] raise exception.PortBindingFailed(port_id=port['id']) [ 814.423095] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] nova.exception.PortBindingFailed: Binding failed for port d62d2305-e5cd-4344-a697-444a752049d9, please check neutron logs for more information. [ 814.423095] env[63021]: ERROR nova.compute.manager [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] [ 814.423095] env[63021]: DEBUG nova.compute.utils [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Binding failed for port d62d2305-e5cd-4344-a697-444a752049d9, please check neutron logs for more information. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 814.427465] env[63021]: DEBUG oslo_concurrency.lockutils [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.862s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.428962] env[63021]: INFO nova.compute.claims [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 814.432017] env[63021]: DEBUG nova.compute.manager [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Build of instance f225812d-f6bf-4350-9151-9ab6c54cce5d was re-scheduled: Binding failed for port d62d2305-e5cd-4344-a697-444a752049d9, please check neutron logs for more information. {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 814.432660] env[63021]: DEBUG nova.compute.manager [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Unplugging VIFs for instance {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 814.432726] env[63021]: DEBUG oslo_concurrency.lockutils [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Acquiring lock "refresh_cache-f225812d-f6bf-4350-9151-9ab6c54cce5d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.432861] env[63021]: DEBUG oslo_concurrency.lockutils [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Acquired lock "refresh_cache-f225812d-f6bf-4350-9151-9ab6c54cce5d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.433018] env[63021]: DEBUG nova.network.neutron [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 814.437862] env[63021]: DEBUG nova.virt.hardware [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 814.438257] env[63021]: DEBUG nova.virt.hardware [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 814.438257] env[63021]: DEBUG nova.virt.hardware [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 814.438412] env[63021]: DEBUG nova.virt.hardware [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 814.438555] env[63021]: DEBUG nova.virt.hardware [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 814.438699] env[63021]: DEBUG nova.virt.hardware [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 814.438926] env[63021]: DEBUG nova.virt.hardware [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 814.439125] env[63021]: DEBUG nova.virt.hardware [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 814.439294] env[63021]: DEBUG nova.virt.hardware [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 814.439452] env[63021]: DEBUG nova.virt.hardware [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 814.439621] env[63021]: DEBUG nova.virt.hardware [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 814.440604] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06a1a8b7-2ad2-4fd2-b59b-4c651dd9a756 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.449619] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88042c11-30e5-4235-a614-cedc5a18f80b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.513699] env[63021]: DEBUG oslo_vmware.api [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293619, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.518915] env[63021]: DEBUG nova.compute.manager [req-206ce1d2-377c-444a-882f-6e9050410088 req-6be98473-a291-4c71-a444-bb79e989b856 service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Received event network-vif-plugged-37dcb3f1-2931-40c6-96ba-b31b8fa1a72b {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 814.519155] env[63021]: DEBUG oslo_concurrency.lockutils [req-206ce1d2-377c-444a-882f-6e9050410088 req-6be98473-a291-4c71-a444-bb79e989b856 service nova] Acquiring lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.519338] env[63021]: DEBUG oslo_concurrency.lockutils [req-206ce1d2-377c-444a-882f-6e9050410088 req-6be98473-a291-4c71-a444-bb79e989b856 service nova] Lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.519508] env[63021]: DEBUG oslo_concurrency.lockutils [req-206ce1d2-377c-444a-882f-6e9050410088 req-6be98473-a291-4c71-a444-bb79e989b856 service nova] Lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.519654] env[63021]: DEBUG nova.compute.manager [req-206ce1d2-377c-444a-882f-6e9050410088 req-6be98473-a291-4c71-a444-bb79e989b856 service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] No waiting events found dispatching network-vif-plugged-37dcb3f1-2931-40c6-96ba-b31b8fa1a72b {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 814.519817] env[63021]: WARNING nova.compute.manager [req-206ce1d2-377c-444a-882f-6e9050410088 req-6be98473-a291-4c71-a444-bb79e989b856 service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Received unexpected event network-vif-plugged-37dcb3f1-2931-40c6-96ba-b31b8fa1a72b for instance with vm_state building and task_state spawning. [ 814.519971] env[63021]: DEBUG nova.compute.manager [req-206ce1d2-377c-444a-882f-6e9050410088 req-6be98473-a291-4c71-a444-bb79e989b856 service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Received event network-changed-37dcb3f1-2931-40c6-96ba-b31b8fa1a72b {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 814.520132] env[63021]: DEBUG nova.compute.manager [req-206ce1d2-377c-444a-882f-6e9050410088 req-6be98473-a291-4c71-a444-bb79e989b856 service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Refreshing instance network info cache due to event network-changed-37dcb3f1-2931-40c6-96ba-b31b8fa1a72b. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 814.520294] env[63021]: DEBUG oslo_concurrency.lockutils [req-206ce1d2-377c-444a-882f-6e9050410088 req-6be98473-a291-4c71-a444-bb79e989b856 service nova] Acquiring lock "refresh_cache-36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.612100] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Releasing lock "refresh_cache-36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.612418] env[63021]: DEBUG nova.compute.manager [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Instance network_info: |[{"id": "37dcb3f1-2931-40c6-96ba-b31b8fa1a72b", "address": "fa:16:3e:57:13:17", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37dcb3f1-29", "ovs_interfaceid": "37dcb3f1-2931-40c6-96ba-b31b8fa1a72b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 814.612708] env[63021]: DEBUG oslo_concurrency.lockutils [req-206ce1d2-377c-444a-882f-6e9050410088 req-6be98473-a291-4c71-a444-bb79e989b856 service nova] Acquired lock "refresh_cache-36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.612879] env[63021]: DEBUG nova.network.neutron [req-206ce1d2-377c-444a-882f-6e9050410088 req-6be98473-a291-4c71-a444-bb79e989b856 service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Refreshing network info cache for port 37dcb3f1-2931-40c6-96ba-b31b8fa1a72b {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 814.614231] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:57:13:17', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8059554c-499f-44b4-be06-29f80ec36b34', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '37dcb3f1-2931-40c6-96ba-b31b8fa1a72b', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 814.621759] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Creating folder: Project (58693d1e991249f28914a39761cc2627). Parent ref: group-v277447. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 814.624498] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-86a77796-0fef-455a-b4a1-56a12456a408 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.635089] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Created folder: Project (58693d1e991249f28914a39761cc2627) in parent group-v277447. [ 814.635280] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Creating folder: Instances. Parent ref: group-v277476. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 814.635504] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a2d8681a-0f62-4a1d-a001-e3d1b80d4537 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.644027] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Created folder: Instances in parent group-v277476. [ 814.644280] env[63021]: DEBUG oslo.service.loopingcall [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 814.644481] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 814.644685] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-22c1f0ee-aa29-49d0-a224-5067b9e41436 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.666088] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 814.666088] env[63021]: value = "task-1293623" [ 814.666088] env[63021]: _type = "Task" [ 814.666088] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.676189] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293623, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.679040] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293620, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.690306] env[63021]: DEBUG oslo_vmware.api [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52411a38-bea5-8277-890c-9c3f70d31326, 'name': SearchDatastore_Task, 'duration_secs': 0.029701} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.691052] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e01b4976-d779-4148-aec0-78f2e16384a2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.696050] env[63021]: DEBUG oslo_vmware.api [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Waiting for the task: (returnval){ [ 814.696050] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52a3f4d7-268b-45b9-a918-4158cb619fd0" [ 814.696050] env[63021]: _type = "Task" [ 814.696050] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.705721] env[63021]: DEBUG oslo_vmware.api [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52a3f4d7-268b-45b9-a918-4158cb619fd0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.810164] env[63021]: INFO nova.compute.manager [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Took 38.63 seconds to build instance. [ 814.835314] env[63021]: DEBUG nova.network.neutron [req-206ce1d2-377c-444a-882f-6e9050410088 req-6be98473-a291-4c71-a444-bb79e989b856 service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Updated VIF entry in instance network info cache for port 37dcb3f1-2931-40c6-96ba-b31b8fa1a72b. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 814.835585] env[63021]: DEBUG nova.network.neutron [req-206ce1d2-377c-444a-882f-6e9050410088 req-6be98473-a291-4c71-a444-bb79e989b856 service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Updating instance_info_cache with network_info: [{"id": "37dcb3f1-2931-40c6-96ba-b31b8fa1a72b", "address": "fa:16:3e:57:13:17", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37dcb3f1-29", "ovs_interfaceid": "37dcb3f1-2931-40c6-96ba-b31b8fa1a72b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.957137] env[63021]: DEBUG nova.network.neutron [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 815.016698] env[63021]: DEBUG oslo_vmware.api [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293619, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.044798] env[63021]: DEBUG nova.network.neutron [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.092150] env[63021]: DEBUG nova.network.neutron [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Successfully updated port: c3780241-4bde-4b91-8bd6-4f42ddc35d07 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 815.179870] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293623, 'name': CreateVM_Task} progress is 25%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.183443] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293620, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.211276] env[63021]: DEBUG oslo_vmware.api [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52a3f4d7-268b-45b9-a918-4158cb619fd0, 'name': SearchDatastore_Task, 'duration_secs': 0.019667} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.213036] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.213036] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] bef5c7cb-be2f-4be7-b72c-4a066d473dd0/bef5c7cb-be2f-4be7-b72c-4a066d473dd0.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 815.213036] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4dd8bb6f-9dc6-4c52-9b33-58e315b2ee23 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.221352] env[63021]: DEBUG oslo_vmware.api [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Waiting for the task: (returnval){ [ 815.221352] env[63021]: value = "task-1293624" [ 815.221352] env[63021]: _type = "Task" [ 815.221352] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.234362] env[63021]: DEBUG oslo_vmware.api [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Task: {'id': task-1293624, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.312949] env[63021]: DEBUG oslo_concurrency.lockutils [None req-09502aa8-375a-43ee-a621-0e431006947b tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Lock "929cf4cd-6f3c-4988-8f80-d889d3070e1e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.954s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.338042] env[63021]: DEBUG oslo_concurrency.lockutils [req-206ce1d2-377c-444a-882f-6e9050410088 req-6be98473-a291-4c71-a444-bb79e989b856 service nova] Releasing lock "refresh_cache-36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.517368] env[63021]: DEBUG oslo_vmware.api [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293619, 'name': ReconfigVM_Task, 'duration_secs': 1.33182} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.517368] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Reconfigured VM instance instance-00000037 to attach disk [datastore1] 7c6d7839-616d-41f5-a909-f8c626477a5e/7c6d7839-616d-41f5-a909-f8c626477a5e.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 815.518672] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8098e7fe-0951-47a2-8c1e-254162b84bfc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.530457] env[63021]: DEBUG oslo_vmware.api [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Waiting for the task: (returnval){ [ 815.530457] env[63021]: value = "task-1293625" [ 815.530457] env[63021]: _type = "Task" [ 815.530457] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.546295] env[63021]: DEBUG oslo_vmware.api [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293625, 'name': Rename_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.548027] env[63021]: DEBUG oslo_concurrency.lockutils [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Releasing lock "refresh_cache-f225812d-f6bf-4350-9151-9ab6c54cce5d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.548193] env[63021]: DEBUG nova.compute.manager [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63021) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 815.548397] env[63021]: DEBUG nova.compute.manager [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 815.548691] env[63021]: DEBUG nova.network.neutron [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 815.576357] env[63021]: DEBUG nova.network.neutron [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 815.594930] env[63021]: DEBUG oslo_concurrency.lockutils [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Acquiring lock "refresh_cache-868b1aeb-8109-4cb9-bb58-33f3d92989ad" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.595129] env[63021]: DEBUG oslo_concurrency.lockutils [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Acquired lock "refresh_cache-868b1aeb-8109-4cb9-bb58-33f3d92989ad" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.595288] env[63021]: DEBUG nova.network.neutron [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 815.681437] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293623, 'name': CreateVM_Task, 'duration_secs': 0.688542} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.686466] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 815.686999] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293620, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.06223} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.687769] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.687929] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.688247] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 815.688496] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 9a6720ef-5307-44b7-b566-8d553c9b5384/9a6720ef-5307-44b7-b566-8d553c9b5384.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 815.688693] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 815.688913] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fba17734-bccb-47ea-81cb-df5a9cd1188a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.690389] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f8c8b7c5-f4c8-47ff-b3ec-aa13d02d62da {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.695400] env[63021]: DEBUG oslo_vmware.api [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 815.695400] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]525e94c0-7517-c6b0-6849-1e6ea0ed331f" [ 815.695400] env[63021]: _type = "Task" [ 815.695400] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.701021] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Waiting for the task: (returnval){ [ 815.701021] env[63021]: value = "task-1293626" [ 815.701021] env[63021]: _type = "Task" [ 815.701021] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.709538] env[63021]: DEBUG oslo_vmware.api [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]525e94c0-7517-c6b0-6849-1e6ea0ed331f, 'name': SearchDatastore_Task, 'duration_secs': 0.008724} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.710206] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.710434] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 815.710657] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.710822] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.711017] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 815.713515] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d193aa9b-ecdb-4160-97da-ea0109c5410e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.718589] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293626, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.726740] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 815.726740] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 815.727284] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a2047b3-171b-45f2-a2f6-bcf0beb47b1e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.734774] env[63021]: DEBUG oslo_vmware.api [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Task: {'id': task-1293624, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.442566} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.735550] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] bef5c7cb-be2f-4be7-b72c-4a066d473dd0/bef5c7cb-be2f-4be7-b72c-4a066d473dd0.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 815.736344] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 815.736344] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-220c8c01-a0f0-4ca4-a85d-f5d3d54b5f9f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.738987] env[63021]: DEBUG oslo_vmware.api [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 815.738987] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5289c077-1f8b-815d-b884-62a85c9dbb07" [ 815.738987] env[63021]: _type = "Task" [ 815.738987] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.743575] env[63021]: DEBUG oslo_vmware.api [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Waiting for the task: (returnval){ [ 815.743575] env[63021]: value = "task-1293627" [ 815.743575] env[63021]: _type = "Task" [ 815.743575] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.754886] env[63021]: DEBUG oslo_vmware.api [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5289c077-1f8b-815d-b884-62a85c9dbb07, 'name': SearchDatastore_Task, 'duration_secs': 0.008494} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.756355] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a3fa02b-4158-4260-93f3-0010a906fa3c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.761684] env[63021]: DEBUG oslo_vmware.api [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Task: {'id': task-1293627, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.765185] env[63021]: DEBUG oslo_vmware.api [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 815.765185] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52734edc-c99b-34cd-e7cc-c8dab7a1f895" [ 815.765185] env[63021]: _type = "Task" [ 815.765185] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.777019] env[63021]: DEBUG oslo_vmware.api [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52734edc-c99b-34cd-e7cc-c8dab7a1f895, 'name': SearchDatastore_Task, 'duration_secs': 0.008125} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.777019] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.777174] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 36ec45ec-761b-4d62-a74f-e4d9a840ada0/36ec45ec-761b-4d62-a74f-e4d9a840ada0.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 815.777424] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1d2208d4-5505-4159-9c73-33f8d0aaf7d3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.783874] env[63021]: DEBUG oslo_vmware.api [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 815.783874] env[63021]: value = "task-1293628" [ 815.783874] env[63021]: _type = "Task" [ 815.783874] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.791624] env[63021]: DEBUG oslo_vmware.api [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293628, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.797680] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-876ac002-be38-47e0-904a-9312148a6484 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.804940] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eee0642-f5df-479d-b4e7-618feb10113a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.834007] env[63021]: DEBUG nova.compute.manager [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 815.840756] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7b9b49f-2f36-496d-a04f-5ae65c949c4f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.847175] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fec9cf6-58d9-4168-a0ae-ebb8ed7fb7f3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.862178] env[63021]: DEBUG nova.compute.provider_tree [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 816.039665] env[63021]: DEBUG oslo_vmware.api [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293625, 'name': Rename_Task, 'duration_secs': 0.219186} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.039946] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 816.040252] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3780357d-661d-443c-b8c5-45ce91646392 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.046333] env[63021]: DEBUG oslo_vmware.api [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Waiting for the task: (returnval){ [ 816.046333] env[63021]: value = "task-1293629" [ 816.046333] env[63021]: _type = "Task" [ 816.046333] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.053201] env[63021]: DEBUG oslo_vmware.api [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293629, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.082098] env[63021]: DEBUG nova.network.neutron [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.141763] env[63021]: DEBUG nova.network.neutron [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 816.170715] env[63021]: DEBUG oslo_concurrency.lockutils [None req-13f4088b-5381-4cac-b728-cd71cf39153c tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Acquiring lock "929cf4cd-6f3c-4988-8f80-d889d3070e1e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.171093] env[63021]: DEBUG oslo_concurrency.lockutils [None req-13f4088b-5381-4cac-b728-cd71cf39153c tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Lock "929cf4cd-6f3c-4988-8f80-d889d3070e1e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.171265] env[63021]: DEBUG oslo_concurrency.lockutils [None req-13f4088b-5381-4cac-b728-cd71cf39153c tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Acquiring lock "929cf4cd-6f3c-4988-8f80-d889d3070e1e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.171517] env[63021]: DEBUG oslo_concurrency.lockutils [None req-13f4088b-5381-4cac-b728-cd71cf39153c tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Lock "929cf4cd-6f3c-4988-8f80-d889d3070e1e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.171738] env[63021]: DEBUG oslo_concurrency.lockutils [None req-13f4088b-5381-4cac-b728-cd71cf39153c tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Lock "929cf4cd-6f3c-4988-8f80-d889d3070e1e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.174167] env[63021]: INFO nova.compute.manager [None req-13f4088b-5381-4cac-b728-cd71cf39153c tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Terminating instance [ 816.176668] env[63021]: DEBUG nova.compute.manager [None req-13f4088b-5381-4cac-b728-cd71cf39153c tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 816.176862] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-13f4088b-5381-4cac-b728-cd71cf39153c tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 816.177811] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-843ded9d-24d8-406d-b0c8-7883143772ce {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.186051] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-13f4088b-5381-4cac-b728-cd71cf39153c tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 816.186188] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9b5ea778-f6c3-4617-b6b4-c6326c1a4745 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.192965] env[63021]: DEBUG oslo_vmware.api [None req-13f4088b-5381-4cac-b728-cd71cf39153c tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Waiting for the task: (returnval){ [ 816.192965] env[63021]: value = "task-1293630" [ 816.192965] env[63021]: _type = "Task" [ 816.192965] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.202491] env[63021]: DEBUG oslo_vmware.api [None req-13f4088b-5381-4cac-b728-cd71cf39153c tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Task: {'id': task-1293630, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.211121] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293626, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070307} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.211469] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 816.212376] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed6a82a9-840a-4025-bcb7-1e70e2cd3ef2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.234169] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Reconfiguring VM instance instance-00000035 to attach disk [datastore1] 9a6720ef-5307-44b7-b566-8d553c9b5384/9a6720ef-5307-44b7-b566-8d553c9b5384.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 816.237074] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cf05ea37-54c6-49b1-85d3-f3382d549e76 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.261070] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Waiting for the task: (returnval){ [ 816.261070] env[63021]: value = "task-1293631" [ 816.261070] env[63021]: _type = "Task" [ 816.261070] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.264171] env[63021]: DEBUG oslo_vmware.api [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Task: {'id': task-1293627, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073596} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.268282] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 816.268420] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4890362-4b75-453a-8860-fbac2f304eaf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.277925] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293631, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.295774] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Reconfiguring VM instance instance-00000038 to attach disk [datastore1] bef5c7cb-be2f-4be7-b72c-4a066d473dd0/bef5c7cb-be2f-4be7-b72c-4a066d473dd0.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 816.302345] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0ed45efb-e16a-440a-9fbf-cdcaa2fcad5b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.324429] env[63021]: DEBUG oslo_vmware.api [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293628, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.326141] env[63021]: DEBUG oslo_vmware.api [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Waiting for the task: (returnval){ [ 816.326141] env[63021]: value = "task-1293632" [ 816.326141] env[63021]: _type = "Task" [ 816.326141] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.336409] env[63021]: DEBUG oslo_vmware.api [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Task: {'id': task-1293632, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.358498] env[63021]: DEBUG oslo_concurrency.lockutils [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.364754] env[63021]: DEBUG nova.scheduler.client.report [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 816.442802] env[63021]: DEBUG nova.network.neutron [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Updating instance_info_cache with network_info: [{"id": "c3780241-4bde-4b91-8bd6-4f42ddc35d07", "address": "fa:16:3e:f3:b2:71", "network": {"id": "dabf10d0-2d5d-4f80-a6c7-9ba893c33efa", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-599669151-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4a0b1326882b40acb72aad7f6c6e2b0a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee555dfd-3d1a-4220-89cd-ffba64e4acf0", "external-id": "nsx-vlan-transportzone-88", "segmentation_id": 88, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc3780241-4b", "ovs_interfaceid": "c3780241-4bde-4b91-8bd6-4f42ddc35d07", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.555602] env[63021]: DEBUG oslo_vmware.api [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293629, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.585444] env[63021]: INFO nova.compute.manager [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] [instance: f225812d-f6bf-4350-9151-9ab6c54cce5d] Took 1.04 seconds to deallocate network for instance. [ 816.594402] env[63021]: DEBUG nova.compute.manager [req-b1b70261-8924-4988-ab44-dc1120117e23 req-af2119b7-4cc4-4134-a83e-369fb0b494b0 service nova] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Received event network-vif-plugged-c3780241-4bde-4b91-8bd6-4f42ddc35d07 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 816.594482] env[63021]: DEBUG oslo_concurrency.lockutils [req-b1b70261-8924-4988-ab44-dc1120117e23 req-af2119b7-4cc4-4134-a83e-369fb0b494b0 service nova] Acquiring lock "868b1aeb-8109-4cb9-bb58-33f3d92989ad-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.594748] env[63021]: DEBUG oslo_concurrency.lockutils [req-b1b70261-8924-4988-ab44-dc1120117e23 req-af2119b7-4cc4-4134-a83e-369fb0b494b0 service nova] Lock "868b1aeb-8109-4cb9-bb58-33f3d92989ad-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.595087] env[63021]: DEBUG oslo_concurrency.lockutils [req-b1b70261-8924-4988-ab44-dc1120117e23 req-af2119b7-4cc4-4134-a83e-369fb0b494b0 service nova] Lock "868b1aeb-8109-4cb9-bb58-33f3d92989ad-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.596299] env[63021]: DEBUG nova.compute.manager [req-b1b70261-8924-4988-ab44-dc1120117e23 req-af2119b7-4cc4-4134-a83e-369fb0b494b0 service nova] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] No waiting events found dispatching network-vif-plugged-c3780241-4bde-4b91-8bd6-4f42ddc35d07 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 816.596299] env[63021]: WARNING nova.compute.manager [req-b1b70261-8924-4988-ab44-dc1120117e23 req-af2119b7-4cc4-4134-a83e-369fb0b494b0 service nova] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Received unexpected event network-vif-plugged-c3780241-4bde-4b91-8bd6-4f42ddc35d07 for instance with vm_state building and task_state spawning. [ 816.596299] env[63021]: DEBUG nova.compute.manager [req-b1b70261-8924-4988-ab44-dc1120117e23 req-af2119b7-4cc4-4134-a83e-369fb0b494b0 service nova] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Received event network-changed-c3780241-4bde-4b91-8bd6-4f42ddc35d07 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 816.596299] env[63021]: DEBUG nova.compute.manager [req-b1b70261-8924-4988-ab44-dc1120117e23 req-af2119b7-4cc4-4134-a83e-369fb0b494b0 service nova] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Refreshing instance network info cache due to event network-changed-c3780241-4bde-4b91-8bd6-4f42ddc35d07. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 816.596299] env[63021]: DEBUG oslo_concurrency.lockutils [req-b1b70261-8924-4988-ab44-dc1120117e23 req-af2119b7-4cc4-4134-a83e-369fb0b494b0 service nova] Acquiring lock "refresh_cache-868b1aeb-8109-4cb9-bb58-33f3d92989ad" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.702627] env[63021]: DEBUG oslo_vmware.api [None req-13f4088b-5381-4cac-b728-cd71cf39153c tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Task: {'id': task-1293630, 'name': PowerOffVM_Task, 'duration_secs': 0.284809} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.702851] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-13f4088b-5381-4cac-b728-cd71cf39153c tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 816.703029] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-13f4088b-5381-4cac-b728-cd71cf39153c tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 816.703279] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-381eea6b-5ac5-4b3a-9002-c747c9ca0de3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.769474] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-13f4088b-5381-4cac-b728-cd71cf39153c tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 816.769685] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-13f4088b-5381-4cac-b728-cd71cf39153c tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 816.769864] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-13f4088b-5381-4cac-b728-cd71cf39153c tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Deleting the datastore file [datastore1] 929cf4cd-6f3c-4988-8f80-d889d3070e1e {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 816.770124] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e96fded6-331c-4462-a201-3ef3fc154b75 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.774677] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293631, 'name': ReconfigVM_Task, 'duration_secs': 0.324651} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.775145] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Reconfigured VM instance instance-00000035 to attach disk [datastore1] 9a6720ef-5307-44b7-b566-8d553c9b5384/9a6720ef-5307-44b7-b566-8d553c9b5384.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 816.775698] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ad2f449a-5afd-43be-8231-916263e78da2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.778705] env[63021]: DEBUG oslo_vmware.api [None req-13f4088b-5381-4cac-b728-cd71cf39153c tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Waiting for the task: (returnval){ [ 816.778705] env[63021]: value = "task-1293634" [ 816.778705] env[63021]: _type = "Task" [ 816.778705] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.782260] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Waiting for the task: (returnval){ [ 816.782260] env[63021]: value = "task-1293635" [ 816.782260] env[63021]: _type = "Task" [ 816.782260] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.788075] env[63021]: DEBUG oslo_vmware.api [None req-13f4088b-5381-4cac-b728-cd71cf39153c tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Task: {'id': task-1293634, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.793785] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293635, 'name': Rename_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.798276] env[63021]: DEBUG oslo_vmware.api [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293628, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.583864} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.798506] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 36ec45ec-761b-4d62-a74f-e4d9a840ada0/36ec45ec-761b-4d62-a74f-e4d9a840ada0.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 816.798711] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 816.798934] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1e4a8c2c-902b-4a74-8ccf-e892fbdcca9d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.805118] env[63021]: DEBUG oslo_vmware.api [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 816.805118] env[63021]: value = "task-1293636" [ 816.805118] env[63021]: _type = "Task" [ 816.805118] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.812237] env[63021]: DEBUG oslo_vmware.api [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293636, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.835149] env[63021]: DEBUG oslo_vmware.api [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Task: {'id': task-1293632, 'name': ReconfigVM_Task, 'duration_secs': 0.302915} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.835461] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Reconfigured VM instance instance-00000038 to attach disk [datastore1] bef5c7cb-be2f-4be7-b72c-4a066d473dd0/bef5c7cb-be2f-4be7-b72c-4a066d473dd0.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 816.836123] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1286143e-cf65-4e34-ad5c-bac673097845 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.841826] env[63021]: DEBUG oslo_vmware.api [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Waiting for the task: (returnval){ [ 816.841826] env[63021]: value = "task-1293637" [ 816.841826] env[63021]: _type = "Task" [ 816.841826] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.849846] env[63021]: DEBUG oslo_vmware.api [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Task: {'id': task-1293637, 'name': Rename_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.870552] env[63021]: DEBUG oslo_concurrency.lockutils [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.443s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.871045] env[63021]: DEBUG nova.compute.manager [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 816.873665] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.547s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.874966] env[63021]: INFO nova.compute.claims [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 816.946251] env[63021]: DEBUG oslo_concurrency.lockutils [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Releasing lock "refresh_cache-868b1aeb-8109-4cb9-bb58-33f3d92989ad" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.946767] env[63021]: DEBUG nova.compute.manager [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Instance network_info: |[{"id": "c3780241-4bde-4b91-8bd6-4f42ddc35d07", "address": "fa:16:3e:f3:b2:71", "network": {"id": "dabf10d0-2d5d-4f80-a6c7-9ba893c33efa", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-599669151-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4a0b1326882b40acb72aad7f6c6e2b0a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee555dfd-3d1a-4220-89cd-ffba64e4acf0", "external-id": "nsx-vlan-transportzone-88", "segmentation_id": 88, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc3780241-4b", "ovs_interfaceid": "c3780241-4bde-4b91-8bd6-4f42ddc35d07", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 816.947142] env[63021]: DEBUG oslo_concurrency.lockutils [req-b1b70261-8924-4988-ab44-dc1120117e23 req-af2119b7-4cc4-4134-a83e-369fb0b494b0 service nova] Acquired lock "refresh_cache-868b1aeb-8109-4cb9-bb58-33f3d92989ad" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.947340] env[63021]: DEBUG nova.network.neutron [req-b1b70261-8924-4988-ab44-dc1120117e23 req-af2119b7-4cc4-4134-a83e-369fb0b494b0 service nova] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Refreshing network info cache for port c3780241-4bde-4b91-8bd6-4f42ddc35d07 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 816.949224] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f3:b2:71', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ee555dfd-3d1a-4220-89cd-ffba64e4acf0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c3780241-4bde-4b91-8bd6-4f42ddc35d07', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 816.958979] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Creating folder: Project (4a0b1326882b40acb72aad7f6c6e2b0a). Parent ref: group-v277447. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 816.960278] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-467b3253-7259-4aa3-9fc2-11824bdfafab {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.971075] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Created folder: Project (4a0b1326882b40acb72aad7f6c6e2b0a) in parent group-v277447. [ 816.971279] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Creating folder: Instances. Parent ref: group-v277479. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 816.971535] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bb1cd157-d53b-4f5d-84f1-aca3d38cb5e1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.980740] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Created folder: Instances in parent group-v277479. [ 816.981018] env[63021]: DEBUG oslo.service.loopingcall [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 816.981216] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 816.981418] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dc3a3269-023b-4707-b474-527e9444aea2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.004371] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 817.004371] env[63021]: value = "task-1293640" [ 817.004371] env[63021]: _type = "Task" [ 817.004371] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.012050] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293640, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.057842] env[63021]: DEBUG oslo_vmware.api [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293629, 'name': PowerOnVM_Task, 'duration_secs': 0.701836} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.057842] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 817.057842] env[63021]: INFO nova.compute.manager [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Took 10.27 seconds to spawn the instance on the hypervisor. [ 817.057842] env[63021]: DEBUG nova.compute.manager [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 817.057842] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-241ec001-c10a-4e41-8667-daa9dbecf012 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.292948] env[63021]: DEBUG oslo_vmware.api [None req-13f4088b-5381-4cac-b728-cd71cf39153c tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Task: {'id': task-1293634, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.296155] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293635, 'name': Rename_Task, 'duration_secs': 0.161449} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.296435] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 817.296699] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bb01c3a8-3753-47df-a83d-ea1e9feb4c31 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.304559] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Waiting for the task: (returnval){ [ 817.304559] env[63021]: value = "task-1293641" [ 817.304559] env[63021]: _type = "Task" [ 817.304559] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.316295] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293641, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.319277] env[63021]: DEBUG oslo_vmware.api [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293636, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065827} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.319416] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 817.320314] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d88a31fc-e154-420d-83e0-6e96020370ab {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.341724] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] 36ec45ec-761b-4d62-a74f-e4d9a840ada0/36ec45ec-761b-4d62-a74f-e4d9a840ada0.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 817.342092] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-696e0d7e-9e52-4ddd-a1ef-9f6a911a35b6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.364465] env[63021]: DEBUG oslo_vmware.api [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Task: {'id': task-1293637, 'name': Rename_Task, 'duration_secs': 0.176296} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.365693] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 817.366008] env[63021]: DEBUG oslo_vmware.api [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 817.366008] env[63021]: value = "task-1293642" [ 817.366008] env[63021]: _type = "Task" [ 817.366008] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.366198] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-685f4f4e-0812-4dda-b580-276a4d39edf2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.376159] env[63021]: DEBUG oslo_vmware.api [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293642, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.377334] env[63021]: DEBUG oslo_vmware.api [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Waiting for the task: (returnval){ [ 817.377334] env[63021]: value = "task-1293643" [ 817.377334] env[63021]: _type = "Task" [ 817.377334] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.378476] env[63021]: DEBUG nova.compute.utils [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 817.384414] env[63021]: DEBUG nova.compute.manager [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 817.384585] env[63021]: DEBUG nova.network.neutron [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 817.391684] env[63021]: DEBUG oslo_vmware.api [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Task: {'id': task-1293643, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.434097] env[63021]: DEBUG nova.policy [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4fe657b588214bc29667fe9453a61038', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2d364c9eae4b4f8a8b60d4a3ac2ed648', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 817.514778] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293640, 'name': CreateVM_Task, 'duration_secs': 0.355635} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.515111] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 817.516362] env[63021]: DEBUG oslo_concurrency.lockutils [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.516362] env[63021]: DEBUG oslo_concurrency.lockutils [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.516362] env[63021]: DEBUG oslo_concurrency.lockutils [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 817.516567] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b601f1b-c543-46a6-bc8b-5e3b3349f568 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.521128] env[63021]: DEBUG oslo_vmware.api [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Waiting for the task: (returnval){ [ 817.521128] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]523ef3a4-c1b0-63b4-0a85-c74b5e70b958" [ 817.521128] env[63021]: _type = "Task" [ 817.521128] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.531078] env[63021]: DEBUG oslo_vmware.api [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]523ef3a4-c1b0-63b4-0a85-c74b5e70b958, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.575637] env[63021]: INFO nova.compute.manager [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Took 39.49 seconds to build instance. [ 817.623219] env[63021]: INFO nova.scheduler.client.report [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Deleted allocations for instance f225812d-f6bf-4350-9151-9ab6c54cce5d [ 817.695483] env[63021]: DEBUG nova.network.neutron [req-b1b70261-8924-4988-ab44-dc1120117e23 req-af2119b7-4cc4-4134-a83e-369fb0b494b0 service nova] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Updated VIF entry in instance network info cache for port c3780241-4bde-4b91-8bd6-4f42ddc35d07. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 817.697098] env[63021]: DEBUG nova.network.neutron [req-b1b70261-8924-4988-ab44-dc1120117e23 req-af2119b7-4cc4-4134-a83e-369fb0b494b0 service nova] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Updating instance_info_cache with network_info: [{"id": "c3780241-4bde-4b91-8bd6-4f42ddc35d07", "address": "fa:16:3e:f3:b2:71", "network": {"id": "dabf10d0-2d5d-4f80-a6c7-9ba893c33efa", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-599669151-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4a0b1326882b40acb72aad7f6c6e2b0a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee555dfd-3d1a-4220-89cd-ffba64e4acf0", "external-id": "nsx-vlan-transportzone-88", "segmentation_id": 88, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc3780241-4b", "ovs_interfaceid": "c3780241-4bde-4b91-8bd6-4f42ddc35d07", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.790750] env[63021]: DEBUG oslo_vmware.api [None req-13f4088b-5381-4cac-b728-cd71cf39153c tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Task: {'id': task-1293634, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.524458} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.791048] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-13f4088b-5381-4cac-b728-cd71cf39153c tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 817.791240] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-13f4088b-5381-4cac-b728-cd71cf39153c tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 817.791411] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-13f4088b-5381-4cac-b728-cd71cf39153c tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 817.791741] env[63021]: INFO nova.compute.manager [None req-13f4088b-5381-4cac-b728-cd71cf39153c tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Took 1.61 seconds to destroy the instance on the hypervisor. [ 817.792039] env[63021]: DEBUG oslo.service.loopingcall [None req-13f4088b-5381-4cac-b728-cd71cf39153c tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 817.792227] env[63021]: DEBUG nova.compute.manager [-] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 817.792320] env[63021]: DEBUG nova.network.neutron [-] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 817.814422] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293641, 'name': PowerOnVM_Task} progress is 90%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.862059] env[63021]: DEBUG nova.network.neutron [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Successfully created port: 948a6d57-d04a-4c5c-ae6f-cd5786ec9a89 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 817.877977] env[63021]: DEBUG oslo_vmware.api [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293642, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.884956] env[63021]: DEBUG nova.compute.manager [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 817.891636] env[63021]: DEBUG oslo_vmware.api [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Task: {'id': task-1293643, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.031955] env[63021]: DEBUG oslo_vmware.api [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]523ef3a4-c1b0-63b4-0a85-c74b5e70b958, 'name': SearchDatastore_Task, 'duration_secs': 0.022947} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.032376] env[63021]: DEBUG oslo_concurrency.lockutils [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.032930] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 818.033235] env[63021]: DEBUG oslo_concurrency.lockutils [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.033383] env[63021]: DEBUG oslo_concurrency.lockutils [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.033568] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 818.033882] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a8a7f762-24cf-4fa3-beea-52a938ceee59 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.047443] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 818.047756] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 818.049980] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd17c071-fe15-46f9-9aa7-1bada4a13c00 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.063140] env[63021]: DEBUG oslo_vmware.api [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Waiting for the task: (returnval){ [ 818.063140] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52c071c4-06b5-0333-fc2e-a1eb25f3d7e5" [ 818.063140] env[63021]: _type = "Task" [ 818.063140] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.076331] env[63021]: DEBUG oslo_vmware.api [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52c071c4-06b5-0333-fc2e-a1eb25f3d7e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.076729] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2207dbe6-ef62-422c-96a4-d04012cd2543 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Lock "7c6d7839-616d-41f5-a909-f8c626477a5e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.398s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.088076] env[63021]: DEBUG nova.compute.manager [req-e8d43e5e-ed9c-48d1-8dab-f4cbbe578a5d req-e95b0839-e01d-4afc-8e5f-12fadf066944 service nova] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Received event network-vif-deleted-034fe798-4091-44e1-9dcd-618125f3a581 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 818.088533] env[63021]: INFO nova.compute.manager [req-e8d43e5e-ed9c-48d1-8dab-f4cbbe578a5d req-e95b0839-e01d-4afc-8e5f-12fadf066944 service nova] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Neutron deleted interface 034fe798-4091-44e1-9dcd-618125f3a581; detaching it from the instance and deleting it from the info cache [ 818.088802] env[63021]: DEBUG nova.network.neutron [req-e8d43e5e-ed9c-48d1-8dab-f4cbbe578a5d req-e95b0839-e01d-4afc-8e5f-12fadf066944 service nova] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.134546] env[63021]: DEBUG oslo_concurrency.lockutils [None req-35253051-0454-4ab8-9579-28e1fc5b16d7 tempest-SecurityGroupsTestJSON-1673128248 tempest-SecurityGroupsTestJSON-1673128248-project-member] Lock "f225812d-f6bf-4350-9151-9ab6c54cce5d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 121.404s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.198564] env[63021]: DEBUG oslo_concurrency.lockutils [req-b1b70261-8924-4988-ab44-dc1120117e23 req-af2119b7-4cc4-4134-a83e-369fb0b494b0 service nova] Releasing lock "refresh_cache-868b1aeb-8109-4cb9-bb58-33f3d92989ad" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.212039] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4af97ff6-c30e-43b4-a9c9-4e61e9ce6aaf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.225770] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d27ee9b1-5dc1-49f7-b059-87c04b5ffcd3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.257560] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54b84e0a-a1ce-488f-b119-e22648b01c57 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.265093] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-259bf61c-e4b6-460c-be6e-2f5112fac9d3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.278331] env[63021]: DEBUG nova.compute.provider_tree [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 818.313975] env[63021]: DEBUG oslo_vmware.api [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293641, 'name': PowerOnVM_Task, 'duration_secs': 0.559255} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.314267] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 818.314461] env[63021]: DEBUG nova.compute.manager [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 818.315287] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5892962-aa5b-4701-ab6b-d22dd79cb3f7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.376987] env[63021]: DEBUG oslo_vmware.api [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293642, 'name': ReconfigVM_Task, 'duration_secs': 0.692634} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.377280] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Reconfigured VM instance instance-00000039 to attach disk [datastore1] 36ec45ec-761b-4d62-a74f-e4d9a840ada0/36ec45ec-761b-4d62-a74f-e4d9a840ada0.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 818.377931] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-80772cd9-00b2-49aa-acfb-ecd422dc3eb3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.386618] env[63021]: DEBUG oslo_vmware.api [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Task: {'id': task-1293643, 'name': PowerOnVM_Task, 'duration_secs': 0.531264} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.387725] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 818.387927] env[63021]: INFO nova.compute.manager [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Took 8.82 seconds to spawn the instance on the hypervisor. [ 818.388127] env[63021]: DEBUG nova.compute.manager [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 818.388473] env[63021]: DEBUG oslo_vmware.api [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 818.388473] env[63021]: value = "task-1293644" [ 818.388473] env[63021]: _type = "Task" [ 818.388473] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.389179] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65f180ae-7e61-4da0-b2ee-780c07ef1d1d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.404397] env[63021]: DEBUG oslo_vmware.api [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293644, 'name': Rename_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.560901] env[63021]: DEBUG nova.network.neutron [-] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.576845] env[63021]: DEBUG oslo_vmware.api [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52c071c4-06b5-0333-fc2e-a1eb25f3d7e5, 'name': SearchDatastore_Task, 'duration_secs': 0.010094} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.578255] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f7c5b87-3a28-42bd-a12e-78491d1eeaad {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.580921] env[63021]: DEBUG nova.compute.manager [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 818.586848] env[63021]: DEBUG oslo_vmware.api [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Waiting for the task: (returnval){ [ 818.586848] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]520e031f-0f3c-0469-4e18-fd17093251b1" [ 818.586848] env[63021]: _type = "Task" [ 818.586848] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.595532] env[63021]: DEBUG oslo_vmware.api [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]520e031f-0f3c-0469-4e18-fd17093251b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.595769] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a4d93f62-725d-4195-8854-c5b346c26290 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.603952] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cbfc14e-c7d0-4a6d-8615-fd61dbdf67b2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.627568] env[63021]: DEBUG nova.compute.manager [req-e8d43e5e-ed9c-48d1-8dab-f4cbbe578a5d req-e95b0839-e01d-4afc-8e5f-12fadf066944 service nova] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Detach interface failed, port_id=034fe798-4091-44e1-9dcd-618125f3a581, reason: Instance 929cf4cd-6f3c-4988-8f80-d889d3070e1e could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 818.638516] env[63021]: DEBUG nova.compute.manager [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 818.785532] env[63021]: DEBUG nova.scheduler.client.report [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 818.836449] env[63021]: DEBUG oslo_concurrency.lockutils [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.901181] env[63021]: DEBUG oslo_vmware.api [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293644, 'name': Rename_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.905398] env[63021]: DEBUG nova.compute.manager [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 818.914034] env[63021]: INFO nova.compute.manager [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Took 38.83 seconds to build instance. [ 818.928236] env[63021]: DEBUG nova.virt.hardware [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 818.929032] env[63021]: DEBUG nova.virt.hardware [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 818.929032] env[63021]: DEBUG nova.virt.hardware [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 818.929032] env[63021]: DEBUG nova.virt.hardware [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 818.929032] env[63021]: DEBUG nova.virt.hardware [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 818.929260] env[63021]: DEBUG nova.virt.hardware [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 818.929507] env[63021]: DEBUG nova.virt.hardware [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 818.929722] env[63021]: DEBUG nova.virt.hardware [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 818.929947] env[63021]: DEBUG nova.virt.hardware [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 818.930188] env[63021]: DEBUG nova.virt.hardware [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 818.930479] env[63021]: DEBUG nova.virt.hardware [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 818.931531] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e1715d4-babe-4d02-9a48-2385491b1ac6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.941255] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8653915d-f06d-452c-8363-063637058c56 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.070037] env[63021]: INFO nova.compute.manager [-] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Took 1.28 seconds to deallocate network for instance. [ 819.100756] env[63021]: DEBUG oslo_vmware.api [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]520e031f-0f3c-0469-4e18-fd17093251b1, 'name': SearchDatastore_Task, 'duration_secs': 0.011782} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.101450] env[63021]: DEBUG oslo_concurrency.lockutils [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.101450] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 868b1aeb-8109-4cb9-bb58-33f3d92989ad/868b1aeb-8109-4cb9-bb58-33f3d92989ad.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 819.101645] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3fe2c39f-2983-4f00-8a63-65a434546a2f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.108679] env[63021]: DEBUG oslo_vmware.api [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Waiting for the task: (returnval){ [ 819.108679] env[63021]: value = "task-1293645" [ 819.108679] env[63021]: _type = "Task" [ 819.108679] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.116476] env[63021]: DEBUG oslo_vmware.api [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Task: {'id': task-1293645, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.161601] env[63021]: DEBUG oslo_concurrency.lockutils [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.193718] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.290880] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.417s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.291624] env[63021]: DEBUG nova.compute.manager [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 819.295168] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.113s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.298884] env[63021]: INFO nova.compute.claims [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 819.403574] env[63021]: DEBUG oslo_vmware.api [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293644, 'name': Rename_Task, 'duration_secs': 0.763959} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.403890] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 819.404157] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-96a34677-fe52-4b85-9f0b-2fe533290163 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.411471] env[63021]: DEBUG oslo_vmware.api [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 819.411471] env[63021]: value = "task-1293646" [ 819.411471] env[63021]: _type = "Task" [ 819.411471] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.417319] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c491b522-5973-46a6-8b99-60f5a70facad tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Lock "bef5c7cb-be2f-4be7-b72c-4a066d473dd0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.954s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.424190] env[63021]: DEBUG oslo_vmware.api [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293646, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.508176] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Acquiring lock "51e4a14c-4792-43cd-9e99-4e06a8f6d6fb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.509086] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Lock "51e4a14c-4792-43cd-9e99-4e06a8f6d6fb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.577971] env[63021]: DEBUG oslo_concurrency.lockutils [None req-13f4088b-5381-4cac-b728-cd71cf39153c tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.622098] env[63021]: DEBUG oslo_vmware.api [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Task: {'id': task-1293645, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.797148] env[63021]: DEBUG nova.compute.manager [req-ce4ad0fe-75d2-480d-9e53-197bb1bbf421 req-dc3116fd-8f0f-46d0-91fe-67588e1b1f13 service nova] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Received event network-vif-plugged-948a6d57-d04a-4c5c-ae6f-cd5786ec9a89 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 819.797148] env[63021]: DEBUG oslo_concurrency.lockutils [req-ce4ad0fe-75d2-480d-9e53-197bb1bbf421 req-dc3116fd-8f0f-46d0-91fe-67588e1b1f13 service nova] Acquiring lock "df0b6974-7f78-4b73-8583-d09754fbacea-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.797148] env[63021]: DEBUG oslo_concurrency.lockutils [req-ce4ad0fe-75d2-480d-9e53-197bb1bbf421 req-dc3116fd-8f0f-46d0-91fe-67588e1b1f13 service nova] Lock "df0b6974-7f78-4b73-8583-d09754fbacea-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.797148] env[63021]: DEBUG oslo_concurrency.lockutils [req-ce4ad0fe-75d2-480d-9e53-197bb1bbf421 req-dc3116fd-8f0f-46d0-91fe-67588e1b1f13 service nova] Lock "df0b6974-7f78-4b73-8583-d09754fbacea-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.797148] env[63021]: DEBUG nova.compute.manager [req-ce4ad0fe-75d2-480d-9e53-197bb1bbf421 req-dc3116fd-8f0f-46d0-91fe-67588e1b1f13 service nova] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] No waiting events found dispatching network-vif-plugged-948a6d57-d04a-4c5c-ae6f-cd5786ec9a89 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 819.797336] env[63021]: WARNING nova.compute.manager [req-ce4ad0fe-75d2-480d-9e53-197bb1bbf421 req-dc3116fd-8f0f-46d0-91fe-67588e1b1f13 service nova] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Received unexpected event network-vif-plugged-948a6d57-d04a-4c5c-ae6f-cd5786ec9a89 for instance with vm_state building and task_state spawning. [ 819.806023] env[63021]: DEBUG nova.compute.utils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 819.809149] env[63021]: DEBUG nova.compute.manager [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 819.809489] env[63021]: DEBUG nova.network.neutron [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 819.851131] env[63021]: INFO nova.compute.manager [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Rebuilding instance [ 819.854159] env[63021]: DEBUG nova.policy [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '124196aaeeb6499dbfdee5d664e4421d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bb745825ae894ffc9ee752a8fbe18279', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 819.897022] env[63021]: DEBUG nova.compute.manager [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 819.897022] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bd98217-3fbb-4ebb-9334-efc4a3ee96aa {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.922786] env[63021]: DEBUG nova.network.neutron [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Successfully updated port: 948a6d57-d04a-4c5c-ae6f-cd5786ec9a89 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 819.929181] env[63021]: DEBUG nova.compute.manager [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 819.931538] env[63021]: DEBUG oslo_vmware.api [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293646, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.127806] env[63021]: DEBUG oslo_vmware.api [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Task: {'id': task-1293645, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.62172} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.127806] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 868b1aeb-8109-4cb9-bb58-33f3d92989ad/868b1aeb-8109-4cb9-bb58-33f3d92989ad.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 820.127806] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 820.128046] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-039103f5-a230-40fa-9797-07f931d50dc0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.134591] env[63021]: DEBUG oslo_vmware.api [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Waiting for the task: (returnval){ [ 820.134591] env[63021]: value = "task-1293647" [ 820.134591] env[63021]: _type = "Task" [ 820.134591] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.142899] env[63021]: DEBUG oslo_vmware.api [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Task: {'id': task-1293647, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.147580] env[63021]: DEBUG nova.network.neutron [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Successfully created port: ad4b54c8-9cd8-4dde-b661-29b32a858e71 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 820.313287] env[63021]: DEBUG nova.compute.manager [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 820.411781] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 820.412215] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7e887517-4978-4036-8a10-5c6cdb797ec0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.426179] env[63021]: DEBUG oslo_concurrency.lockutils [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquiring lock "refresh_cache-df0b6974-7f78-4b73-8583-d09754fbacea" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.426325] env[63021]: DEBUG oslo_concurrency.lockutils [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquired lock "refresh_cache-df0b6974-7f78-4b73-8583-d09754fbacea" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.426544] env[63021]: DEBUG nova.network.neutron [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 820.431380] env[63021]: DEBUG oslo_vmware.api [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293646, 'name': PowerOnVM_Task, 'duration_secs': 0.602} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.431380] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 820.431380] env[63021]: INFO nova.compute.manager [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Took 8.43 seconds to spawn the instance on the hypervisor. [ 820.431380] env[63021]: DEBUG nova.compute.manager [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 820.431380] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Waiting for the task: (returnval){ [ 820.431380] env[63021]: value = "task-1293648" [ 820.431380] env[63021]: _type = "Task" [ 820.431380] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.432034] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3969669-73a8-410e-95d5-3fa49eee4798 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.455858] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Task: {'id': task-1293648, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.457375] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.644699] env[63021]: DEBUG oslo_vmware.api [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Task: {'id': task-1293647, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.261953} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.647947] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 820.649009] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e901fbe9-7a63-4a00-bdfc-fea443a1a6af {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.673419] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] 868b1aeb-8109-4cb9-bb58-33f3d92989ad/868b1aeb-8109-4cb9-bb58-33f3d92989ad.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 820.675995] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3ef830b4-dccb-43b2-b3e9-cd4657e2f1f1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.695990] env[63021]: DEBUG oslo_vmware.api [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Waiting for the task: (returnval){ [ 820.695990] env[63021]: value = "task-1293649" [ 820.695990] env[63021]: _type = "Task" [ 820.695990] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.703716] env[63021]: DEBUG oslo_vmware.api [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Task: {'id': task-1293649, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.716718] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c0d5fcb-56f9-4fe1-acbd-a0156d06f13f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.724461] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c96f17af-c480-4010-98fc-5968c833b148 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.756112] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22561621-4537-47df-af37-2c1e732219f8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.763479] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-347ebe06-1bda-4862-9ab3-96b0f4591fd6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.776832] env[63021]: DEBUG nova.compute.provider_tree [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Updating inventory in ProviderTree for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 820.945670] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Task: {'id': task-1293648, 'name': PowerOffVM_Task, 'duration_secs': 0.182865} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.945922] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 820.946155] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 820.947229] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5725e3e-5c6e-4498-8f1b-8851c97dcd86 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.955028] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 820.955171] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3378e95f-6a0e-4c2b-a451-5af5b7fb41de {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.965583] env[63021]: INFO nova.compute.manager [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Took 39.11 seconds to build instance. [ 820.977454] env[63021]: DEBUG nova.network.neutron [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 820.984209] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 820.984477] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 820.984698] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Deleting the datastore file [datastore1] 9a6720ef-5307-44b7-b566-8d553c9b5384 {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 820.985408] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-432039f0-e714-40db-86c2-ce5b752777a0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.991432] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Waiting for the task: (returnval){ [ 820.991432] env[63021]: value = "task-1293651" [ 820.991432] env[63021]: _type = "Task" [ 820.991432] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.000416] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Task: {'id': task-1293651, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.201636] env[63021]: DEBUG nova.network.neutron [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Updating instance_info_cache with network_info: [{"id": "948a6d57-d04a-4c5c-ae6f-cd5786ec9a89", "address": "fa:16:3e:b6:53:e3", "network": {"id": "37008b1f-68ff-404c-9afc-14b108af85c3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1473182647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "2d364c9eae4b4f8a8b60d4a3ac2ed648", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "02bbcead-d833-4543-bec6-fb82dfe659ff", "external-id": "nsx-vlan-transportzone-478", "segmentation_id": 478, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap948a6d57-d0", "ovs_interfaceid": "948a6d57-d04a-4c5c-ae6f-cd5786ec9a89", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.207331] env[63021]: DEBUG oslo_vmware.api [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Task: {'id': task-1293649, 'name': ReconfigVM_Task, 'duration_secs': 0.473952} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.207331] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Reconfigured VM instance instance-0000003a to attach disk [datastore2] 868b1aeb-8109-4cb9-bb58-33f3d92989ad/868b1aeb-8109-4cb9-bb58-33f3d92989ad.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 821.207774] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-335f0bf0-29d9-4511-aa9a-53e363f7278c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.217226] env[63021]: DEBUG oslo_vmware.api [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Waiting for the task: (returnval){ [ 821.217226] env[63021]: value = "task-1293652" [ 821.217226] env[63021]: _type = "Task" [ 821.217226] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.229259] env[63021]: DEBUG oslo_vmware.api [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Task: {'id': task-1293652, 'name': Rename_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.300247] env[63021]: ERROR nova.scheduler.client.report [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [req-a01121fd-7610-4fc6-92c1-c2249998b45b] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 6ef6c041-2b0d-4bc3-8662-5d679a77a86d. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-a01121fd-7610-4fc6-92c1-c2249998b45b"}]} [ 821.321218] env[63021]: DEBUG nova.scheduler.client.report [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Refreshing inventories for resource provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 821.324377] env[63021]: DEBUG nova.compute.manager [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 821.343629] env[63021]: DEBUG nova.scheduler.client.report [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Updating ProviderTree inventory for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 821.343943] env[63021]: DEBUG nova.compute.provider_tree [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Updating inventory in ProviderTree for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 821.357893] env[63021]: DEBUG nova.virt.hardware [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 821.358136] env[63021]: DEBUG nova.virt.hardware [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 821.358291] env[63021]: DEBUG nova.virt.hardware [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 821.358534] env[63021]: DEBUG nova.virt.hardware [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 821.358645] env[63021]: DEBUG nova.virt.hardware [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 821.358801] env[63021]: DEBUG nova.virt.hardware [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 821.359019] env[63021]: DEBUG nova.virt.hardware [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 821.359554] env[63021]: DEBUG nova.virt.hardware [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 821.359735] env[63021]: DEBUG nova.virt.hardware [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 821.359902] env[63021]: DEBUG nova.virt.hardware [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 821.360130] env[63021]: DEBUG nova.virt.hardware [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 821.362381] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80678f33-10fd-449d-a953-73d8e89da0b2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.368652] env[63021]: DEBUG nova.scheduler.client.report [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Refreshing aggregate associations for resource provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d, aggregates: None {{(pid=63021) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 821.374732] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-689efa00-c0c0-4618-b4f3-b5dac45e6b31 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.395058] env[63021]: DEBUG nova.scheduler.client.report [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Refreshing trait associations for resource provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE {{(pid=63021) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 821.468883] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c1c8c4e3-dfdf-4591-b545-a7b0b1a4348b tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 98.170s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.501807] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Task: {'id': task-1293651, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138553} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.505621] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 821.505896] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 821.506127] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 821.707431] env[63021]: DEBUG oslo_concurrency.lockutils [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Releasing lock "refresh_cache-df0b6974-7f78-4b73-8583-d09754fbacea" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.707884] env[63021]: DEBUG nova.compute.manager [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Instance network_info: |[{"id": "948a6d57-d04a-4c5c-ae6f-cd5786ec9a89", "address": "fa:16:3e:b6:53:e3", "network": {"id": "37008b1f-68ff-404c-9afc-14b108af85c3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1473182647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "2d364c9eae4b4f8a8b60d4a3ac2ed648", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "02bbcead-d833-4543-bec6-fb82dfe659ff", "external-id": "nsx-vlan-transportzone-478", "segmentation_id": 478, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap948a6d57-d0", "ovs_interfaceid": "948a6d57-d04a-4c5c-ae6f-cd5786ec9a89", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 821.708359] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b6:53:e3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '02bbcead-d833-4543-bec6-fb82dfe659ff', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '948a6d57-d04a-4c5c-ae6f-cd5786ec9a89', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 821.715916] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Creating folder: Project (2d364c9eae4b4f8a8b60d4a3ac2ed648). Parent ref: group-v277447. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 821.716239] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-af90ae09-a1fb-4474-aca0-4024abe4fad9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.731113] env[63021]: DEBUG oslo_vmware.api [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Task: {'id': task-1293652, 'name': Rename_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.733848] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Created folder: Project (2d364c9eae4b4f8a8b60d4a3ac2ed648) in parent group-v277447. [ 821.734007] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Creating folder: Instances. Parent ref: group-v277482. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 821.735138] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2606edaa-24f2-49eb-a959-9a19900e3afb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.737173] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf9574c4-4285-4aff-a511-b4fcab2652e2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.745530] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2316afd0-9f99-4062-b3b1-d60f6dcf8d22 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.749578] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Created folder: Instances in parent group-v277482. [ 821.749806] env[63021]: DEBUG oslo.service.loopingcall [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 821.750330] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 821.750527] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-35ebf93f-c372-4e4e-b707-e08f9e023c87 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.795647] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45afd386-f6bd-45fc-8364-3ab99874f618 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.801456] env[63021]: DEBUG nova.network.neutron [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Successfully updated port: ad4b54c8-9cd8-4dde-b661-29b32a858e71 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 821.801456] env[63021]: DEBUG oslo_concurrency.lockutils [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Acquiring lock "bef5c7cb-be2f-4be7-b72c-4a066d473dd0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.801521] env[63021]: DEBUG oslo_concurrency.lockutils [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Lock "bef5c7cb-be2f-4be7-b72c-4a066d473dd0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.801808] env[63021]: DEBUG oslo_concurrency.lockutils [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Acquiring lock "bef5c7cb-be2f-4be7-b72c-4a066d473dd0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.801888] env[63021]: DEBUG oslo_concurrency.lockutils [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Lock "bef5c7cb-be2f-4be7-b72c-4a066d473dd0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.802059] env[63021]: DEBUG oslo_concurrency.lockutils [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Lock "bef5c7cb-be2f-4be7-b72c-4a066d473dd0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.803559] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 821.803559] env[63021]: value = "task-1293655" [ 821.803559] env[63021]: _type = "Task" [ 821.803559] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.805962] env[63021]: DEBUG nova.compute.manager [req-473cd768-3dfe-4782-8f1f-79b69ddfb5e9 req-f60e1de2-a20b-479c-b23c-6994764095cb service nova] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Received event network-vif-plugged-ad4b54c8-9cd8-4dde-b661-29b32a858e71 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 821.806276] env[63021]: DEBUG oslo_concurrency.lockutils [req-473cd768-3dfe-4782-8f1f-79b69ddfb5e9 req-f60e1de2-a20b-479c-b23c-6994764095cb service nova] Acquiring lock "2209d709-7a5b-42e2-be93-b4fd436b6290-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.806386] env[63021]: DEBUG oslo_concurrency.lockutils [req-473cd768-3dfe-4782-8f1f-79b69ddfb5e9 req-f60e1de2-a20b-479c-b23c-6994764095cb service nova] Lock "2209d709-7a5b-42e2-be93-b4fd436b6290-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.806552] env[63021]: DEBUG oslo_concurrency.lockutils [req-473cd768-3dfe-4782-8f1f-79b69ddfb5e9 req-f60e1de2-a20b-479c-b23c-6994764095cb service nova] Lock "2209d709-7a5b-42e2-be93-b4fd436b6290-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.806712] env[63021]: DEBUG nova.compute.manager [req-473cd768-3dfe-4782-8f1f-79b69ddfb5e9 req-f60e1de2-a20b-479c-b23c-6994764095cb service nova] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] No waiting events found dispatching network-vif-plugged-ad4b54c8-9cd8-4dde-b661-29b32a858e71 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 821.806917] env[63021]: WARNING nova.compute.manager [req-473cd768-3dfe-4782-8f1f-79b69ddfb5e9 req-f60e1de2-a20b-479c-b23c-6994764095cb service nova] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Received unexpected event network-vif-plugged-ad4b54c8-9cd8-4dde-b661-29b32a858e71 for instance with vm_state building and task_state spawning. [ 821.807355] env[63021]: INFO nova.compute.manager [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Terminating instance [ 821.818465] env[63021]: DEBUG nova.compute.manager [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 821.818465] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 821.818465] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fce81a21-eee7-4f18-8481-bd129f177c90 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.821138] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3694f752-6445-476d-b3a7-45bdb9e23da7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.830861] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293655, 'name': CreateVM_Task} progress is 15%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.841284] env[63021]: DEBUG nova.compute.provider_tree [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Updating inventory in ProviderTree for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 821.846754] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 821.847207] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-327135c0-1609-46a9-9cda-9ab63e37c3b5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.853237] env[63021]: DEBUG oslo_vmware.api [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Waiting for the task: (returnval){ [ 821.853237] env[63021]: value = "task-1293656" [ 821.853237] env[63021]: _type = "Task" [ 821.853237] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.861370] env[63021]: DEBUG oslo_vmware.api [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Task: {'id': task-1293656, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.966399] env[63021]: DEBUG nova.compute.manager [req-3a81aabf-87cd-45ca-98e4-eb8710739794 req-06be42a4-7ade-4e2f-9ee0-01b0db695411 service nova] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Received event network-changed-948a6d57-d04a-4c5c-ae6f-cd5786ec9a89 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 821.966512] env[63021]: DEBUG nova.compute.manager [req-3a81aabf-87cd-45ca-98e4-eb8710739794 req-06be42a4-7ade-4e2f-9ee0-01b0db695411 service nova] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Refreshing instance network info cache due to event network-changed-948a6d57-d04a-4c5c-ae6f-cd5786ec9a89. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 821.966719] env[63021]: DEBUG oslo_concurrency.lockutils [req-3a81aabf-87cd-45ca-98e4-eb8710739794 req-06be42a4-7ade-4e2f-9ee0-01b0db695411 service nova] Acquiring lock "refresh_cache-df0b6974-7f78-4b73-8583-d09754fbacea" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.967740] env[63021]: DEBUG oslo_concurrency.lockutils [req-3a81aabf-87cd-45ca-98e4-eb8710739794 req-06be42a4-7ade-4e2f-9ee0-01b0db695411 service nova] Acquired lock "refresh_cache-df0b6974-7f78-4b73-8583-d09754fbacea" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.969206] env[63021]: DEBUG nova.network.neutron [req-3a81aabf-87cd-45ca-98e4-eb8710739794 req-06be42a4-7ade-4e2f-9ee0-01b0db695411 service nova] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Refreshing network info cache for port 948a6d57-d04a-4c5c-ae6f-cd5786ec9a89 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 821.976369] env[63021]: DEBUG nova.compute.manager [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 822.234020] env[63021]: DEBUG oslo_vmware.api [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Task: {'id': task-1293652, 'name': Rename_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.304867] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquiring lock "refresh_cache-2209d709-7a5b-42e2-be93-b4fd436b6290" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.305051] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquired lock "refresh_cache-2209d709-7a5b-42e2-be93-b4fd436b6290" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.305217] env[63021]: DEBUG nova.network.neutron [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 822.322834] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293655, 'name': CreateVM_Task} progress is 25%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.362565] env[63021]: DEBUG oslo_vmware.api [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Task: {'id': task-1293656, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.398505] env[63021]: DEBUG nova.scheduler.client.report [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Updated inventory for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with generation 83 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 822.398783] env[63021]: DEBUG nova.compute.provider_tree [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Updating resource provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d generation from 83 to 84 during operation: update_inventory {{(pid=63021) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 822.398965] env[63021]: DEBUG nova.compute.provider_tree [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Updating inventory in ProviderTree for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 822.505439] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.547605] env[63021]: DEBUG nova.virt.hardware [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 822.547883] env[63021]: DEBUG nova.virt.hardware [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 822.547994] env[63021]: DEBUG nova.virt.hardware [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 822.548174] env[63021]: DEBUG nova.virt.hardware [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 822.548363] env[63021]: DEBUG nova.virt.hardware [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 822.551185] env[63021]: DEBUG nova.virt.hardware [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 822.551185] env[63021]: DEBUG nova.virt.hardware [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 822.551185] env[63021]: DEBUG nova.virt.hardware [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 822.551185] env[63021]: DEBUG nova.virt.hardware [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 822.551185] env[63021]: DEBUG nova.virt.hardware [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 822.551765] env[63021]: DEBUG nova.virt.hardware [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 822.551765] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff9fe1d3-2bd0-45eb-acdb-a1000a615034 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.561882] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb75621b-9ef0-4dcf-baf3-284d1b2f6ce2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.578614] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Instance VIF info [] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 822.591024] env[63021]: DEBUG oslo.service.loopingcall [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 822.591153] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 822.593399] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-63af0b03-2c75-497e-89d1-3f0601c0c744 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.612580] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 822.612580] env[63021]: value = "task-1293657" [ 822.612580] env[63021]: _type = "Task" [ 822.612580] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.622758] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293657, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.734825] env[63021]: DEBUG oslo_vmware.api [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Task: {'id': task-1293652, 'name': Rename_Task, 'duration_secs': 1.451103} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.739134] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 822.739134] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2da8fc42-f466-4084-a5ab-d7923c64977b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.743931] env[63021]: DEBUG oslo_vmware.api [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Waiting for the task: (returnval){ [ 822.743931] env[63021]: value = "task-1293658" [ 822.743931] env[63021]: _type = "Task" [ 822.743931] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.745044] env[63021]: DEBUG nova.network.neutron [req-3a81aabf-87cd-45ca-98e4-eb8710739794 req-06be42a4-7ade-4e2f-9ee0-01b0db695411 service nova] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Updated VIF entry in instance network info cache for port 948a6d57-d04a-4c5c-ae6f-cd5786ec9a89. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 822.747017] env[63021]: DEBUG nova.network.neutron [req-3a81aabf-87cd-45ca-98e4-eb8710739794 req-06be42a4-7ade-4e2f-9ee0-01b0db695411 service nova] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Updating instance_info_cache with network_info: [{"id": "948a6d57-d04a-4c5c-ae6f-cd5786ec9a89", "address": "fa:16:3e:b6:53:e3", "network": {"id": "37008b1f-68ff-404c-9afc-14b108af85c3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1473182647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "2d364c9eae4b4f8a8b60d4a3ac2ed648", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "02bbcead-d833-4543-bec6-fb82dfe659ff", "external-id": "nsx-vlan-transportzone-478", "segmentation_id": 478, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap948a6d57-d0", "ovs_interfaceid": "948a6d57-d04a-4c5c-ae6f-cd5786ec9a89", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.755489] env[63021]: DEBUG oslo_vmware.api [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Task: {'id': task-1293658, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.822081] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293655, 'name': CreateVM_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.864142] env[63021]: DEBUG oslo_vmware.api [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Task: {'id': task-1293656, 'name': PowerOffVM_Task, 'duration_secs': 1.007221} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.864913] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 822.865428] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 822.866103] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b9b6446c-a1c7-41d0-97e2-b9142850b47d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.881160] env[63021]: DEBUG nova.network.neutron [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 822.912022] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.614s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.912022] env[63021]: DEBUG nova.compute.manager [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 822.913212] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 12.874s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.929080] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 822.929080] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 822.929080] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Deleting the datastore file [datastore1] bef5c7cb-be2f-4be7-b72c-4a066d473dd0 {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 822.929080] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a0008805-236d-4d87-9b56-870062db7adc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.939625] env[63021]: DEBUG oslo_vmware.api [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Waiting for the task: (returnval){ [ 822.939625] env[63021]: value = "task-1293660" [ 822.939625] env[63021]: _type = "Task" [ 822.939625] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.949119] env[63021]: DEBUG oslo_vmware.api [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Task: {'id': task-1293660, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.081327] env[63021]: DEBUG nova.network.neutron [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Updating instance_info_cache with network_info: [{"id": "ad4b54c8-9cd8-4dde-b661-29b32a858e71", "address": "fa:16:3e:99:e5:0a", "network": {"id": "111ca5aa-2e4a-498d-b78a-81a1e2ad9295", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2056436219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb745825ae894ffc9ee752a8fbe18279", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad4b54c8-9c", "ovs_interfaceid": "ad4b54c8-9cd8-4dde-b661-29b32a858e71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.122362] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293657, 'name': CreateVM_Task, 'duration_secs': 0.320529} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.122537] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 823.122986] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.123212] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.123571] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 823.124766] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b3a71df-4ed6-4dc1-924e-cdef95d36e55 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.128675] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Waiting for the task: (returnval){ [ 823.128675] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52a318ca-f735-f8d5-61e6-646f3fd37043" [ 823.128675] env[63021]: _type = "Task" [ 823.128675] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.139483] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52a318ca-f735-f8d5-61e6-646f3fd37043, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.252251] env[63021]: DEBUG oslo_concurrency.lockutils [req-3a81aabf-87cd-45ca-98e4-eb8710739794 req-06be42a4-7ade-4e2f-9ee0-01b0db695411 service nova] Releasing lock "refresh_cache-df0b6974-7f78-4b73-8583-d09754fbacea" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.257949] env[63021]: DEBUG oslo_vmware.api [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Task: {'id': task-1293658, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.325433] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293655, 'name': CreateVM_Task, 'duration_secs': 1.105547} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.325683] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 823.326602] env[63021]: DEBUG oslo_concurrency.lockutils [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.421043] env[63021]: DEBUG nova.compute.utils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 823.422620] env[63021]: DEBUG nova.compute.manager [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 823.422869] env[63021]: DEBUG nova.network.neutron [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 823.454221] env[63021]: DEBUG oslo_vmware.api [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Task: {'id': task-1293660, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.194308} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.454343] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 823.454506] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 823.454757] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 823.454994] env[63021]: INFO nova.compute.manager [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Took 1.64 seconds to destroy the instance on the hypervisor. [ 823.455334] env[63021]: DEBUG oslo.service.loopingcall [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 823.455581] env[63021]: DEBUG nova.compute.manager [-] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 823.455708] env[63021]: DEBUG nova.network.neutron [-] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 823.480450] env[63021]: DEBUG nova.policy [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '124196aaeeb6499dbfdee5d664e4421d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bb745825ae894ffc9ee752a8fbe18279', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 823.584050] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Releasing lock "refresh_cache-2209d709-7a5b-42e2-be93-b4fd436b6290" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.584392] env[63021]: DEBUG nova.compute.manager [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Instance network_info: |[{"id": "ad4b54c8-9cd8-4dde-b661-29b32a858e71", "address": "fa:16:3e:99:e5:0a", "network": {"id": "111ca5aa-2e4a-498d-b78a-81a1e2ad9295", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2056436219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb745825ae894ffc9ee752a8fbe18279", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad4b54c8-9c", "ovs_interfaceid": "ad4b54c8-9cd8-4dde-b661-29b32a858e71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 823.584841] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:99:e5:0a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4c9a12d2-469f-4199-bfaa-f791d765deac', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ad4b54c8-9cd8-4dde-b661-29b32a858e71', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 823.592494] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Creating folder: Project (bb745825ae894ffc9ee752a8fbe18279). Parent ref: group-v277447. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 823.593107] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-44a44194-6a08-4716-9c10-f79867f17c3e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.603198] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Created folder: Project (bb745825ae894ffc9ee752a8fbe18279) in parent group-v277447. [ 823.603395] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Creating folder: Instances. Parent ref: group-v277486. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 823.603630] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-444d5ed6-bb74-4f5d-9cf1-a5edce93503f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.614169] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Created folder: Instances in parent group-v277486. [ 823.614400] env[63021]: DEBUG oslo.service.loopingcall [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 823.614577] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 823.614769] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5984bc54-9417-46db-aaea-1715686f4e2e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.633386] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 823.633386] env[63021]: value = "task-1293663" [ 823.633386] env[63021]: _type = "Task" [ 823.633386] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.640883] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52a318ca-f735-f8d5-61e6-646f3fd37043, 'name': SearchDatastore_Task, 'duration_secs': 0.010705} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.643108] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.643108] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 823.643108] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.643108] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.643250] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 823.647526] env[63021]: DEBUG oslo_concurrency.lockutils [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.648164] env[63021]: DEBUG oslo_concurrency.lockutils [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 823.648435] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e13b89fc-0994-4b77-91be-27d6a89f1dbe {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.651182] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293663, 'name': CreateVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.651393] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59598c20-5fc5-4cfa-9832-2cdec433e0be {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.656358] env[63021]: DEBUG oslo_vmware.api [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 823.656358] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5229ed33-1130-2705-637c-19118ef7b2d2" [ 823.656358] env[63021]: _type = "Task" [ 823.656358] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.661754] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 823.661989] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 823.663278] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b516a7b-dd0a-477d-9106-1212a03b8035 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.673404] env[63021]: DEBUG oslo_vmware.api [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5229ed33-1130-2705-637c-19118ef7b2d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.675664] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Waiting for the task: (returnval){ [ 823.675664] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52f40d9d-d310-3de5-a7ca-5cee49c13cbe" [ 823.675664] env[63021]: _type = "Task" [ 823.675664] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.688472] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52f40d9d-d310-3de5-a7ca-5cee49c13cbe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.758175] env[63021]: DEBUG oslo_vmware.api [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Task: {'id': task-1293658, 'name': PowerOnVM_Task, 'duration_secs': 0.52822} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.758455] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 823.758577] env[63021]: INFO nova.compute.manager [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Took 9.35 seconds to spawn the instance on the hypervisor. [ 823.758753] env[63021]: DEBUG nova.compute.manager [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 823.759565] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd77d649-7a52-4cb1-8743-9dfb767b61e6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.771462] env[63021]: DEBUG nova.network.neutron [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Successfully created port: 8fee9441-e591-4a58-8296-3ec48ab9c154 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 823.908550] env[63021]: DEBUG nova.compute.manager [req-eb915839-a028-4c45-9e74-c6d145ff7e1a req-f60b04d2-80f3-473d-9794-431183497dd8 service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Received event network-changed-37dcb3f1-2931-40c6-96ba-b31b8fa1a72b {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 823.908843] env[63021]: DEBUG nova.compute.manager [req-eb915839-a028-4c45-9e74-c6d145ff7e1a req-f60b04d2-80f3-473d-9794-431183497dd8 service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Refreshing instance network info cache due to event network-changed-37dcb3f1-2931-40c6-96ba-b31b8fa1a72b. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 823.909132] env[63021]: DEBUG oslo_concurrency.lockutils [req-eb915839-a028-4c45-9e74-c6d145ff7e1a req-f60b04d2-80f3-473d-9794-431183497dd8 service nova] Acquiring lock "refresh_cache-36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.909344] env[63021]: DEBUG oslo_concurrency.lockutils [req-eb915839-a028-4c45-9e74-c6d145ff7e1a req-f60b04d2-80f3-473d-9794-431183497dd8 service nova] Acquired lock "refresh_cache-36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.909553] env[63021]: DEBUG nova.network.neutron [req-eb915839-a028-4c45-9e74-c6d145ff7e1a req-f60b04d2-80f3-473d-9794-431183497dd8 service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Refreshing network info cache for port 37dcb3f1-2931-40c6-96ba-b31b8fa1a72b {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 823.928828] env[63021]: DEBUG nova.compute.manager [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 823.956602] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 9a6720ef-5307-44b7-b566-8d553c9b5384 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 823.956892] env[63021]: WARNING nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 929cf4cd-6f3c-4988-8f80-d889d3070e1e is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 823.957072] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 7c6d7839-616d-41f5-a909-f8c626477a5e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 823.957204] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance bef5c7cb-be2f-4be7-b72c-4a066d473dd0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 823.957424] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 36ec45ec-761b-4d62-a74f-e4d9a840ada0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 823.957481] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 868b1aeb-8109-4cb9-bb58-33f3d92989ad actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 823.957555] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance df0b6974-7f78-4b73-8583-d09754fbacea actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 823.957664] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 2209d709-7a5b-42e2-be93-b4fd436b6290 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 823.957805] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 36d26354-6127-4557-acb0-8743e27ff1ff actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 824.076069] env[63021]: DEBUG nova.compute.manager [req-b8bd6099-90c7-4bb7-9bf6-20b410fcef2d req-a4360c85-4174-4aa7-b313-4f105843f9d4 service nova] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Received event network-vif-deleted-7b37c85f-8b13-40cf-9407-1c618513578c {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 824.076164] env[63021]: INFO nova.compute.manager [req-b8bd6099-90c7-4bb7-9bf6-20b410fcef2d req-a4360c85-4174-4aa7-b313-4f105843f9d4 service nova] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Neutron deleted interface 7b37c85f-8b13-40cf-9407-1c618513578c; detaching it from the instance and deleting it from the info cache [ 824.076311] env[63021]: DEBUG nova.network.neutron [req-b8bd6099-90c7-4bb7-9bf6-20b410fcef2d req-a4360c85-4174-4aa7-b313-4f105843f9d4 service nova] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.145336] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293663, 'name': CreateVM_Task, 'duration_secs': 0.36629} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.146378] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 824.146378] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.169714] env[63021]: DEBUG oslo_vmware.api [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5229ed33-1130-2705-637c-19118ef7b2d2, 'name': SearchDatastore_Task, 'duration_secs': 0.013751} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.170321] env[63021]: DEBUG oslo_concurrency.lockutils [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.171204] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 824.171204] env[63021]: DEBUG oslo_concurrency.lockutils [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.171533] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.171858] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 824.172465] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bbaddd58-58b4-447d-83c6-4f2ac8439ea4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.182629] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Waiting for the task: (returnval){ [ 824.182629] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52cc87f6-cd93-6175-c336-79afe40b2735" [ 824.182629] env[63021]: _type = "Task" [ 824.182629] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.193252] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52f40d9d-d310-3de5-a7ca-5cee49c13cbe, 'name': SearchDatastore_Task, 'duration_secs': 0.018444} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.194401] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-890842b1-26bf-4458-bc92-1dcc27d2e3c6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.199880] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52cc87f6-cd93-6175-c336-79afe40b2735, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.202825] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Waiting for the task: (returnval){ [ 824.202825] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]520c22ff-e9d0-435c-8b32-8882b77b0223" [ 824.202825] env[63021]: _type = "Task" [ 824.202825] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.210884] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]520c22ff-e9d0-435c-8b32-8882b77b0223, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.275132] env[63021]: DEBUG nova.network.neutron [-] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.279448] env[63021]: INFO nova.compute.manager [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Took 30.08 seconds to build instance. [ 824.463585] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 824.579546] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fdea5071-e60f-416c-946b-7572475f7618 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.591494] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eabe715-5489-45bc-afd0-8e4101609be8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.626622] env[63021]: DEBUG nova.compute.manager [req-b8bd6099-90c7-4bb7-9bf6-20b410fcef2d req-a4360c85-4174-4aa7-b313-4f105843f9d4 service nova] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Detach interface failed, port_id=7b37c85f-8b13-40cf-9407-1c618513578c, reason: Instance bef5c7cb-be2f-4be7-b72c-4a066d473dd0 could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 824.696032] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52cc87f6-cd93-6175-c336-79afe40b2735, 'name': SearchDatastore_Task, 'duration_secs': 0.029254} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.696413] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.696725] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 824.696989] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.713694] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]520c22ff-e9d0-435c-8b32-8882b77b0223, 'name': SearchDatastore_Task, 'duration_secs': 0.026103} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.714335] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.714689] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 9a6720ef-5307-44b7-b566-8d553c9b5384/9a6720ef-5307-44b7-b566-8d553c9b5384.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 824.715359] env[63021]: DEBUG oslo_concurrency.lockutils [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.715359] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 824.715576] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7d520cfe-b1c1-4c1c-8059-7d6667b75a3f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.718138] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5d87aa79-d374-4c26-96d6-709c71ecba3b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.724837] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Waiting for the task: (returnval){ [ 824.724837] env[63021]: value = "task-1293664" [ 824.724837] env[63021]: _type = "Task" [ 824.724837] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.730137] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 824.730373] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 824.731465] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57032e11-ab0f-43bb-be72-d2f1b48869ce {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.739768] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Task: {'id': task-1293664, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.743398] env[63021]: DEBUG oslo_vmware.api [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 824.743398] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5225686a-1c69-11d4-8e26-d7f2af0158b6" [ 824.743398] env[63021]: _type = "Task" [ 824.743398] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.753660] env[63021]: DEBUG oslo_vmware.api [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5225686a-1c69-11d4-8e26-d7f2af0158b6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.779967] env[63021]: INFO nova.compute.manager [-] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Took 1.32 seconds to deallocate network for instance. [ 824.790024] env[63021]: DEBUG oslo_concurrency.lockutils [None req-46258eef-9141-4c2f-a59e-fb9f82bc6695 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Lock "868b1aeb-8109-4cb9-bb58-33f3d92989ad" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 91.642s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.932158] env[63021]: DEBUG nova.network.neutron [req-eb915839-a028-4c45-9e74-c6d145ff7e1a req-f60b04d2-80f3-473d-9794-431183497dd8 service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Updated VIF entry in instance network info cache for port 37dcb3f1-2931-40c6-96ba-b31b8fa1a72b. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 824.932746] env[63021]: DEBUG nova.network.neutron [req-eb915839-a028-4c45-9e74-c6d145ff7e1a req-f60b04d2-80f3-473d-9794-431183497dd8 service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Updating instance_info_cache with network_info: [{"id": "37dcb3f1-2931-40c6-96ba-b31b8fa1a72b", "address": "fa:16:3e:57:13:17", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37dcb3f1-29", "ovs_interfaceid": "37dcb3f1-2931-40c6-96ba-b31b8fa1a72b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.944184] env[63021]: DEBUG nova.compute.manager [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 824.968226] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 6f4e6245-c0d9-46b4-baf5-5aca19e66da5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 824.981268] env[63021]: DEBUG nova.virt.hardware [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 824.981754] env[63021]: DEBUG nova.virt.hardware [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 824.981886] env[63021]: DEBUG nova.virt.hardware [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 824.982040] env[63021]: DEBUG nova.virt.hardware [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 824.982204] env[63021]: DEBUG nova.virt.hardware [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 824.982388] env[63021]: DEBUG nova.virt.hardware [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 824.982726] env[63021]: DEBUG nova.virt.hardware [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 824.982819] env[63021]: DEBUG nova.virt.hardware [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 824.983022] env[63021]: DEBUG nova.virt.hardware [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 824.983704] env[63021]: DEBUG nova.virt.hardware [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 824.983704] env[63021]: DEBUG nova.virt.hardware [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 824.984474] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fb4156f-b12f-4766-b2e2-93ef9b2b7799 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.995881] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b015b9d6-ee4e-4e7e-9adc-51efa47ee08f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.237960] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Task: {'id': task-1293664, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.261181] env[63021]: DEBUG oslo_vmware.api [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5225686a-1c69-11d4-8e26-d7f2af0158b6, 'name': SearchDatastore_Task, 'duration_secs': 0.010454} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.261181] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6bb565e4-be69-41e0-a194-764bd4be68f2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.270168] env[63021]: DEBUG oslo_vmware.api [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 825.270168] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]520988ac-5f9b-95ee-215e-189c50b6eec4" [ 825.270168] env[63021]: _type = "Task" [ 825.270168] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.281405] env[63021]: DEBUG oslo_vmware.api [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]520988ac-5f9b-95ee-215e-189c50b6eec4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.288599] env[63021]: DEBUG oslo_concurrency.lockutils [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.292813] env[63021]: DEBUG nova.compute.manager [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 825.438434] env[63021]: DEBUG oslo_concurrency.lockutils [req-eb915839-a028-4c45-9e74-c6d145ff7e1a req-f60b04d2-80f3-473d-9794-431183497dd8 service nova] Releasing lock "refresh_cache-36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.438434] env[63021]: DEBUG nova.compute.manager [req-eb915839-a028-4c45-9e74-c6d145ff7e1a req-f60b04d2-80f3-473d-9794-431183497dd8 service nova] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Received event network-changed-ad4b54c8-9cd8-4dde-b661-29b32a858e71 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 825.438434] env[63021]: DEBUG nova.compute.manager [req-eb915839-a028-4c45-9e74-c6d145ff7e1a req-f60b04d2-80f3-473d-9794-431183497dd8 service nova] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Refreshing instance network info cache due to event network-changed-ad4b54c8-9cd8-4dde-b661-29b32a858e71. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 825.438434] env[63021]: DEBUG oslo_concurrency.lockutils [req-eb915839-a028-4c45-9e74-c6d145ff7e1a req-f60b04d2-80f3-473d-9794-431183497dd8 service nova] Acquiring lock "refresh_cache-2209d709-7a5b-42e2-be93-b4fd436b6290" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.438434] env[63021]: DEBUG oslo_concurrency.lockutils [req-eb915839-a028-4c45-9e74-c6d145ff7e1a req-f60b04d2-80f3-473d-9794-431183497dd8 service nova] Acquired lock "refresh_cache-2209d709-7a5b-42e2-be93-b4fd436b6290" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.438636] env[63021]: DEBUG nova.network.neutron [req-eb915839-a028-4c45-9e74-c6d145ff7e1a req-f60b04d2-80f3-473d-9794-431183497dd8 service nova] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Refreshing network info cache for port ad4b54c8-9cd8-4dde-b661-29b32a858e71 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 825.473548] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 70f67905-7c60-433d-9ebc-d66fa44eb36e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 825.590755] env[63021]: DEBUG nova.network.neutron [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Successfully updated port: 8fee9441-e591-4a58-8296-3ec48ab9c154 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 825.739049] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Task: {'id': task-1293664, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.712728} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.739230] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 9a6720ef-5307-44b7-b566-8d553c9b5384/9a6720ef-5307-44b7-b566-8d553c9b5384.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 825.739462] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 825.739725] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-946b9772-a4f9-43b5-a513-49141b48f086 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.746745] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Waiting for the task: (returnval){ [ 825.746745] env[63021]: value = "task-1293665" [ 825.746745] env[63021]: _type = "Task" [ 825.746745] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.754219] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Task: {'id': task-1293665, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.778955] env[63021]: DEBUG oslo_vmware.api [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]520988ac-5f9b-95ee-215e-189c50b6eec4, 'name': SearchDatastore_Task, 'duration_secs': 0.055011} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.779305] env[63021]: DEBUG oslo_concurrency.lockutils [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.779469] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] df0b6974-7f78-4b73-8583-d09754fbacea/df0b6974-7f78-4b73-8583-d09754fbacea.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 825.779850] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.779942] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 825.780182] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-94241316-36ea-4e2f-9ba4-38aa463ec1b2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.782701] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2d8b5bad-5332-45e3-b1aa-5274c7d8012e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.789296] env[63021]: DEBUG oslo_vmware.api [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 825.789296] env[63021]: value = "task-1293666" [ 825.789296] env[63021]: _type = "Task" [ 825.789296] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.793484] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 825.793665] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 825.797506] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a527092-5b92-46ab-bc94-3f36798f11b6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.805444] env[63021]: DEBUG oslo_vmware.api [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293666, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.808833] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Waiting for the task: (returnval){ [ 825.808833] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]524c8424-7ae5-08a5-7742-427334784169" [ 825.808833] env[63021]: _type = "Task" [ 825.808833] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.816946] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]524c8424-7ae5-08a5-7742-427334784169, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.822635] env[63021]: DEBUG oslo_concurrency.lockutils [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.957289] env[63021]: DEBUG nova.compute.manager [req-e8a4e1f3-943f-4e77-af0c-1baa0e077b23 req-d516432a-663f-4e7a-9730-106c91680cf5 service nova] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Received event network-vif-plugged-8fee9441-e591-4a58-8296-3ec48ab9c154 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 825.957289] env[63021]: DEBUG oslo_concurrency.lockutils [req-e8a4e1f3-943f-4e77-af0c-1baa0e077b23 req-d516432a-663f-4e7a-9730-106c91680cf5 service nova] Acquiring lock "36d26354-6127-4557-acb0-8743e27ff1ff-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.957289] env[63021]: DEBUG oslo_concurrency.lockutils [req-e8a4e1f3-943f-4e77-af0c-1baa0e077b23 req-d516432a-663f-4e7a-9730-106c91680cf5 service nova] Lock "36d26354-6127-4557-acb0-8743e27ff1ff-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.957289] env[63021]: DEBUG oslo_concurrency.lockutils [req-e8a4e1f3-943f-4e77-af0c-1baa0e077b23 req-d516432a-663f-4e7a-9730-106c91680cf5 service nova] Lock "36d26354-6127-4557-acb0-8743e27ff1ff-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.957289] env[63021]: DEBUG nova.compute.manager [req-e8a4e1f3-943f-4e77-af0c-1baa0e077b23 req-d516432a-663f-4e7a-9730-106c91680cf5 service nova] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] No waiting events found dispatching network-vif-plugged-8fee9441-e591-4a58-8296-3ec48ab9c154 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 825.957542] env[63021]: WARNING nova.compute.manager [req-e8a4e1f3-943f-4e77-af0c-1baa0e077b23 req-d516432a-663f-4e7a-9730-106c91680cf5 service nova] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Received unexpected event network-vif-plugged-8fee9441-e591-4a58-8296-3ec48ab9c154 for instance with vm_state building and task_state spawning. [ 825.957542] env[63021]: DEBUG nova.compute.manager [req-e8a4e1f3-943f-4e77-af0c-1baa0e077b23 req-d516432a-663f-4e7a-9730-106c91680cf5 service nova] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Received event network-changed-8fee9441-e591-4a58-8296-3ec48ab9c154 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 825.957542] env[63021]: DEBUG nova.compute.manager [req-e8a4e1f3-943f-4e77-af0c-1baa0e077b23 req-d516432a-663f-4e7a-9730-106c91680cf5 service nova] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Refreshing instance network info cache due to event network-changed-8fee9441-e591-4a58-8296-3ec48ab9c154. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 825.957542] env[63021]: DEBUG oslo_concurrency.lockutils [req-e8a4e1f3-943f-4e77-af0c-1baa0e077b23 req-d516432a-663f-4e7a-9730-106c91680cf5 service nova] Acquiring lock "refresh_cache-36d26354-6127-4557-acb0-8743e27ff1ff" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.957542] env[63021]: DEBUG oslo_concurrency.lockutils [req-e8a4e1f3-943f-4e77-af0c-1baa0e077b23 req-d516432a-663f-4e7a-9730-106c91680cf5 service nova] Acquired lock "refresh_cache-36d26354-6127-4557-acb0-8743e27ff1ff" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.957769] env[63021]: DEBUG nova.network.neutron [req-e8a4e1f3-943f-4e77-af0c-1baa0e077b23 req-d516432a-663f-4e7a-9730-106c91680cf5 service nova] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Refreshing network info cache for port 8fee9441-e591-4a58-8296-3ec48ab9c154 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 825.979863] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 7787f6bb-e855-476f-9146-2a648932429d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 826.094257] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquiring lock "refresh_cache-36d26354-6127-4557-acb0-8743e27ff1ff" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.245652] env[63021]: DEBUG nova.network.neutron [req-eb915839-a028-4c45-9e74-c6d145ff7e1a req-f60b04d2-80f3-473d-9794-431183497dd8 service nova] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Updated VIF entry in instance network info cache for port ad4b54c8-9cd8-4dde-b661-29b32a858e71. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 826.245652] env[63021]: DEBUG nova.network.neutron [req-eb915839-a028-4c45-9e74-c6d145ff7e1a req-f60b04d2-80f3-473d-9794-431183497dd8 service nova] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Updating instance_info_cache with network_info: [{"id": "ad4b54c8-9cd8-4dde-b661-29b32a858e71", "address": "fa:16:3e:99:e5:0a", "network": {"id": "111ca5aa-2e4a-498d-b78a-81a1e2ad9295", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2056436219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb745825ae894ffc9ee752a8fbe18279", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad4b54c8-9c", "ovs_interfaceid": "ad4b54c8-9cd8-4dde-b661-29b32a858e71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.257221] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Task: {'id': task-1293665, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071515} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.258166] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 826.259029] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb537840-72dd-4a48-95b4-e757ca85a6ee {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.280503] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Reconfiguring VM instance instance-00000035 to attach disk [datastore2] 9a6720ef-5307-44b7-b566-8d553c9b5384/9a6720ef-5307-44b7-b566-8d553c9b5384.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 826.281288] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fc6e2f10-4bd9-4d68-8262-c8c5d603b2da {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.307509] env[63021]: DEBUG oslo_vmware.api [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293666, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.309051] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Waiting for the task: (returnval){ [ 826.309051] env[63021]: value = "task-1293667" [ 826.309051] env[63021]: _type = "Task" [ 826.309051] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.321230] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Task: {'id': task-1293667, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.325254] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]524c8424-7ae5-08a5-7742-427334784169, 'name': SearchDatastore_Task, 'duration_secs': 0.009967} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.326174] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6526ca23-5f8f-4aac-9fb3-28db93198359 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.336030] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Waiting for the task: (returnval){ [ 826.336030] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]523a28c5-cb0b-93a0-e49f-261f80453b39" [ 826.336030] env[63021]: _type = "Task" [ 826.336030] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.343192] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]523a28c5-cb0b-93a0-e49f-261f80453b39, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.484421] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 71909213-258d-45f5-9a3e-e473e8ce8aa2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 826.501951] env[63021]: DEBUG nova.network.neutron [req-e8a4e1f3-943f-4e77-af0c-1baa0e077b23 req-d516432a-663f-4e7a-9730-106c91680cf5 service nova] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 826.615102] env[63021]: DEBUG nova.network.neutron [req-e8a4e1f3-943f-4e77-af0c-1baa0e077b23 req-d516432a-663f-4e7a-9730-106c91680cf5 service nova] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.747604] env[63021]: DEBUG oslo_concurrency.lockutils [req-eb915839-a028-4c45-9e74-c6d145ff7e1a req-f60b04d2-80f3-473d-9794-431183497dd8 service nova] Releasing lock "refresh_cache-2209d709-7a5b-42e2-be93-b4fd436b6290" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.809413] env[63021]: DEBUG oslo_vmware.api [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293666, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.624089} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.810343] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] df0b6974-7f78-4b73-8583-d09754fbacea/df0b6974-7f78-4b73-8583-d09754fbacea.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 826.812101] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 826.814110] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-13f32080-6771-4845-bcc2-5fb1f38f3ead {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.821287] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Task: {'id': task-1293667, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.822693] env[63021]: DEBUG oslo_vmware.api [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 826.822693] env[63021]: value = "task-1293668" [ 826.822693] env[63021]: _type = "Task" [ 826.822693] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.831062] env[63021]: DEBUG oslo_vmware.api [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293668, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.842435] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]523a28c5-cb0b-93a0-e49f-261f80453b39, 'name': SearchDatastore_Task, 'duration_secs': 0.035145} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.842638] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.842889] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 2209d709-7a5b-42e2-be93-b4fd436b6290/2209d709-7a5b-42e2-be93-b4fd436b6290.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 826.843150] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-08de34f0-b62d-470a-a549-95ecfda2aa21 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.849023] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Waiting for the task: (returnval){ [ 826.849023] env[63021]: value = "task-1293669" [ 826.849023] env[63021]: _type = "Task" [ 826.849023] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.855532] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': task-1293669, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.988275] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance df856146-61ba-4de7-a5fd-d75263927163 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 827.117865] env[63021]: DEBUG oslo_concurrency.lockutils [req-e8a4e1f3-943f-4e77-af0c-1baa0e077b23 req-d516432a-663f-4e7a-9730-106c91680cf5 service nova] Releasing lock "refresh_cache-36d26354-6127-4557-acb0-8743e27ff1ff" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.118292] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquired lock "refresh_cache-36d26354-6127-4557-acb0-8743e27ff1ff" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.118446] env[63021]: DEBUG nova.network.neutron [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 827.323635] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Task: {'id': task-1293667, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.331845] env[63021]: DEBUG oslo_vmware.api [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293668, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068375} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.332167] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 827.332890] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4a30aae-ab3f-4451-883b-6aa3fae2c1b7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.355797] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] df0b6974-7f78-4b73-8583-d09754fbacea/df0b6974-7f78-4b73-8583-d09754fbacea.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 827.356352] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cb1d131f-5c4b-4fe6-a3eb-2a99bd6e4dce {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.383659] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': task-1293669, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.384367] env[63021]: DEBUG oslo_vmware.api [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 827.384367] env[63021]: value = "task-1293670" [ 827.384367] env[63021]: _type = "Task" [ 827.384367] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.392906] env[63021]: DEBUG oslo_vmware.api [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293670, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.493305] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 45c9e2be-eb19-4b83-b280-c9eeaddfccba has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 827.601899] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Acquiring lock "868b1aeb-8109-4cb9-bb58-33f3d92989ad" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.602273] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Lock "868b1aeb-8109-4cb9-bb58-33f3d92989ad" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.602506] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Acquiring lock "868b1aeb-8109-4cb9-bb58-33f3d92989ad-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.602742] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Lock "868b1aeb-8109-4cb9-bb58-33f3d92989ad-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.602941] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Lock "868b1aeb-8109-4cb9-bb58-33f3d92989ad-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.605359] env[63021]: INFO nova.compute.manager [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Terminating instance [ 827.607587] env[63021]: DEBUG nova.compute.manager [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 827.607797] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 827.608666] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-251b1f96-6a21-4d7d-8418-8e82014568f9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.617011] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 827.617323] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5f1fc608-32d7-4619-b994-a076f03ab3d3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.624019] env[63021]: DEBUG oslo_vmware.api [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Waiting for the task: (returnval){ [ 827.624019] env[63021]: value = "task-1293671" [ 827.624019] env[63021]: _type = "Task" [ 827.624019] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.634390] env[63021]: DEBUG oslo_vmware.api [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Task: {'id': task-1293671, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.662899] env[63021]: DEBUG nova.network.neutron [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 827.821083] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Task: {'id': task-1293667, 'name': ReconfigVM_Task, 'duration_secs': 1.18318} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.824080] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Reconfigured VM instance instance-00000035 to attach disk [datastore2] 9a6720ef-5307-44b7-b566-8d553c9b5384/9a6720ef-5307-44b7-b566-8d553c9b5384.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 827.824080] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f51c482b-dcc8-4779-93be-b62966e6e678 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.829452] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Waiting for the task: (returnval){ [ 827.829452] env[63021]: value = "task-1293675" [ 827.829452] env[63021]: _type = "Task" [ 827.829452] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.836367] env[63021]: DEBUG nova.network.neutron [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Updating instance_info_cache with network_info: [{"id": "8fee9441-e591-4a58-8296-3ec48ab9c154", "address": "fa:16:3e:f3:82:03", "network": {"id": "111ca5aa-2e4a-498d-b78a-81a1e2ad9295", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2056436219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb745825ae894ffc9ee752a8fbe18279", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fee9441-e5", "ovs_interfaceid": "8fee9441-e591-4a58-8296-3ec48ab9c154", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.842798] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Task: {'id': task-1293675, 'name': Rename_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.868053] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': task-1293669, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.958414} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.868053] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 2209d709-7a5b-42e2-be93-b4fd436b6290/2209d709-7a5b-42e2-be93-b4fd436b6290.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 827.868053] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 827.868053] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dde6091e-870d-44d4-af05-799910b09c94 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.874795] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Waiting for the task: (returnval){ [ 827.874795] env[63021]: value = "task-1293676" [ 827.874795] env[63021]: _type = "Task" [ 827.874795] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.884794] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': task-1293676, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.893113] env[63021]: DEBUG oslo_vmware.api [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293670, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.998028] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance c328d2f7-3398-4f25-b11c-f464be7af8a1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 828.135321] env[63021]: DEBUG oslo_vmware.api [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Task: {'id': task-1293671, 'name': PowerOffVM_Task, 'duration_secs': 0.286669} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.135600] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 828.135793] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 828.136076] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-851ed207-d1b1-42b5-acea-ba0d387a1295 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.193802] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 828.194076] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Deleting contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 828.194351] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Deleting the datastore file [datastore2] 868b1aeb-8109-4cb9-bb58-33f3d92989ad {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 828.194542] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0ba1aac3-0d60-4e2d-84b3-5919e629d754 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.201234] env[63021]: DEBUG oslo_vmware.api [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Waiting for the task: (returnval){ [ 828.201234] env[63021]: value = "task-1293678" [ 828.201234] env[63021]: _type = "Task" [ 828.201234] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.209074] env[63021]: DEBUG oslo_vmware.api [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Task: {'id': task-1293678, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.339883] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Task: {'id': task-1293675, 'name': Rename_Task, 'duration_secs': 0.166046} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.340035] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 828.340410] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dbcb09c0-8f0e-4289-ad2e-fdd3a0c4917a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.342170] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Releasing lock "refresh_cache-36d26354-6127-4557-acb0-8743e27ff1ff" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.342470] env[63021]: DEBUG nova.compute.manager [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Instance network_info: |[{"id": "8fee9441-e591-4a58-8296-3ec48ab9c154", "address": "fa:16:3e:f3:82:03", "network": {"id": "111ca5aa-2e4a-498d-b78a-81a1e2ad9295", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2056436219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb745825ae894ffc9ee752a8fbe18279", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fee9441-e5", "ovs_interfaceid": "8fee9441-e591-4a58-8296-3ec48ab9c154", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 828.343299] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f3:82:03', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4c9a12d2-469f-4199-bfaa-f791d765deac', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8fee9441-e591-4a58-8296-3ec48ab9c154', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 828.351497] env[63021]: DEBUG oslo.service.loopingcall [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 828.352941] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 828.353270] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Waiting for the task: (returnval){ [ 828.353270] env[63021]: value = "task-1293679" [ 828.353270] env[63021]: _type = "Task" [ 828.353270] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.353452] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-96b71868-33b8-4c63-8228-be833906b279 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.383306] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 828.383306] env[63021]: value = "task-1293680" [ 828.383306] env[63021]: _type = "Task" [ 828.383306] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.384135] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Task: {'id': task-1293679, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.391224] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': task-1293676, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068677} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.394782] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 828.395677] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8e934ca-dc68-4c84-aec9-e28529f2f5c1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.401729] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293680, 'name': CreateVM_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.407427] env[63021]: DEBUG oslo_vmware.api [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293670, 'name': ReconfigVM_Task, 'duration_secs': 0.591094} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.417667] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Reconfigured VM instance instance-0000003b to attach disk [datastore2] df0b6974-7f78-4b73-8583-d09754fbacea/df0b6974-7f78-4b73-8583-d09754fbacea.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 828.429318] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] 2209d709-7a5b-42e2-be93-b4fd436b6290/2209d709-7a5b-42e2-be93-b4fd436b6290.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 828.429648] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2ad121e9-16df-4ebd-bee1-f8c3b64138b3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.431500] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2c65f436-2676-4337-b1fb-43522b18cb81 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.451897] env[63021]: DEBUG oslo_vmware.api [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 828.451897] env[63021]: value = "task-1293681" [ 828.451897] env[63021]: _type = "Task" [ 828.451897] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.453353] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Waiting for the task: (returnval){ [ 828.453353] env[63021]: value = "task-1293682" [ 828.453353] env[63021]: _type = "Task" [ 828.453353] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.465584] env[63021]: DEBUG oslo_vmware.api [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293681, 'name': Rename_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.469238] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': task-1293682, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.504638] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 16b21798-17e8-4aeb-affa-57eae31d5dd3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 828.711122] env[63021]: DEBUG oslo_vmware.api [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Task: {'id': task-1293678, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.346051} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.711477] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 828.711683] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Deleted contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 828.711951] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 828.712211] env[63021]: INFO nova.compute.manager [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Took 1.10 seconds to destroy the instance on the hypervisor. [ 828.712476] env[63021]: DEBUG oslo.service.loopingcall [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 828.712737] env[63021]: DEBUG nova.compute.manager [-] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 828.712864] env[63021]: DEBUG nova.network.neutron [-] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 828.878041] env[63021]: DEBUG oslo_vmware.api [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Task: {'id': task-1293679, 'name': PowerOnVM_Task, 'duration_secs': 0.477721} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.878459] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 828.878459] env[63021]: DEBUG nova.compute.manager [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 828.879426] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0955d547-8c87-4167-97b8-0d00e1454753 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.894559] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293680, 'name': CreateVM_Task, 'duration_secs': 0.383208} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.894744] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 828.895376] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.895543] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.895839] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 828.896102] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-882fed2b-22bc-485c-8606-f45225ee62a7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.901716] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Waiting for the task: (returnval){ [ 828.901716] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52fe88be-c23e-3fce-3dec-303b3eff11b6" [ 828.901716] env[63021]: _type = "Task" [ 828.901716] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.908689] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52fe88be-c23e-3fce-3dec-303b3eff11b6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.966327] env[63021]: DEBUG oslo_vmware.api [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293681, 'name': Rename_Task, 'duration_secs': 0.145718} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.968969] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 828.968969] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': task-1293682, 'name': ReconfigVM_Task, 'duration_secs': 0.320768} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.970261] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3a365eeb-c30a-4ce4-9cbb-3c890bf0a170 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.971812] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Reconfigured VM instance instance-0000003c to attach disk [datastore2] 2209d709-7a5b-42e2-be93-b4fd436b6290/2209d709-7a5b-42e2-be93-b4fd436b6290.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 828.972625] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-84ef6a96-73a7-4eb1-a342-0b27301252d6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.979700] env[63021]: DEBUG oslo_vmware.api [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 828.979700] env[63021]: value = "task-1293683" [ 828.979700] env[63021]: _type = "Task" [ 828.979700] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.980874] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Waiting for the task: (returnval){ [ 828.980874] env[63021]: value = "task-1293684" [ 828.980874] env[63021]: _type = "Task" [ 828.980874] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.997104] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': task-1293684, 'name': Rename_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.000721] env[63021]: DEBUG oslo_vmware.api [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293683, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.002491] env[63021]: DEBUG nova.compute.manager [req-461175e4-9a35-46bd-bb69-0fe380f2a3af req-082d9d59-9e37-4cff-ac78-d2ba88d9917f service nova] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Received event network-vif-deleted-c3780241-4bde-4b91-8bd6-4f42ddc35d07 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 829.002784] env[63021]: INFO nova.compute.manager [req-461175e4-9a35-46bd-bb69-0fe380f2a3af req-082d9d59-9e37-4cff-ac78-d2ba88d9917f service nova] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Neutron deleted interface c3780241-4bde-4b91-8bd6-4f42ddc35d07; detaching it from the instance and deleting it from the info cache [ 829.003548] env[63021]: DEBUG nova.network.neutron [req-461175e4-9a35-46bd-bb69-0fe380f2a3af req-082d9d59-9e37-4cff-ac78-d2ba88d9917f service nova] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.008618] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 881c2fb8-5f8c-48c2-a173-8fab949f24f7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 829.398849] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.410415] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52fe88be-c23e-3fce-3dec-303b3eff11b6, 'name': SearchDatastore_Task, 'duration_secs': 0.012859} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.410705] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.410933] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 829.411173] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.411314] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.411488] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 829.411746] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c2482f7a-57c7-4640-8f2c-5d785fbfaa4e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.419857] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 829.420059] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 829.420731] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c8b4b88-63b3-4391-92f8-b616851d652f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.425893] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Waiting for the task: (returnval){ [ 829.425893] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52d07119-9d33-371a-8054-77d48a075ab6" [ 829.425893] env[63021]: _type = "Task" [ 829.425893] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.433530] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52d07119-9d33-371a-8054-77d48a075ab6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.473078] env[63021]: DEBUG nova.network.neutron [-] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.493451] env[63021]: DEBUG oslo_vmware.api [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293683, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.496305] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': task-1293684, 'name': Rename_Task, 'duration_secs': 0.302842} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.496550] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 829.496783] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fc4feb87-91e3-4e69-ae01-82fb907890a0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.503013] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Waiting for the task: (returnval){ [ 829.503013] env[63021]: value = "task-1293685" [ 829.503013] env[63021]: _type = "Task" [ 829.503013] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.511035] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance f0154308-e6a8-4321-a082-99291344664e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 829.512230] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': task-1293685, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.512366] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b37bac83-ae7c-4727-b249-95edcdc7ad8f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.519886] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b129a94-bccf-4ae3-b027-02b98cacd9ea {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.547367] env[63021]: DEBUG nova.compute.manager [req-461175e4-9a35-46bd-bb69-0fe380f2a3af req-082d9d59-9e37-4cff-ac78-d2ba88d9917f service nova] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Detach interface failed, port_id=c3780241-4bde-4b91-8bd6-4f42ddc35d07, reason: Instance 868b1aeb-8109-4cb9-bb58-33f3d92989ad could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 829.938239] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52d07119-9d33-371a-8054-77d48a075ab6, 'name': SearchDatastore_Task, 'duration_secs': 0.017506} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.939051] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d36f0c68-4ab2-4742-923c-b5dbc136d115 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.944641] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Waiting for the task: (returnval){ [ 829.944641] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52107144-ef8b-37c7-513a-3a971e49b1d7" [ 829.944641] env[63021]: _type = "Task" [ 829.944641] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.952700] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52107144-ef8b-37c7-513a-3a971e49b1d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.975875] env[63021]: INFO nova.compute.manager [-] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Took 1.26 seconds to deallocate network for instance. [ 829.993205] env[63021]: DEBUG oslo_vmware.api [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293683, 'name': PowerOnVM_Task, 'duration_secs': 0.915511} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.993647] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 829.993919] env[63021]: INFO nova.compute.manager [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Took 11.09 seconds to spawn the instance on the hypervisor. [ 829.994115] env[63021]: DEBUG nova.compute.manager [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 829.995328] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ba16454-59b8-4027-9864-c9de319fdc3e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.017799] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 44dadf8e-a727-448f-887d-6408f7ebfbc6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 830.025116] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': task-1293685, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.456874] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52107144-ef8b-37c7-513a-3a971e49b1d7, 'name': SearchDatastore_Task, 'duration_secs': 0.010296} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.457069] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.457295] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 36d26354-6127-4557-acb0-8743e27ff1ff/36d26354-6127-4557-acb0-8743e27ff1ff.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 830.457572] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a4168a68-b358-4cf1-a159-646e993b729c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.466438] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Waiting for the task: (returnval){ [ 830.466438] env[63021]: value = "task-1293686" [ 830.466438] env[63021]: _type = "Task" [ 830.466438] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.475743] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': task-1293686, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.487908] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.513090] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': task-1293685, 'name': PowerOnVM_Task, 'duration_secs': 0.70957} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.513363] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 830.513559] env[63021]: INFO nova.compute.manager [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Took 9.19 seconds to spawn the instance on the hypervisor. [ 830.513729] env[63021]: DEBUG nova.compute.manager [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 830.514541] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b288f4e-741c-404e-9e49-eae3350fc662 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.526587] env[63021]: INFO nova.compute.manager [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Took 32.10 seconds to build instance. [ 830.527689] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 830.551601] env[63021]: DEBUG oslo_concurrency.lockutils [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Acquiring lock "9a6720ef-5307-44b7-b566-8d553c9b5384" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.551962] env[63021]: DEBUG oslo_concurrency.lockutils [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Lock "9a6720ef-5307-44b7-b566-8d553c9b5384" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.552239] env[63021]: DEBUG oslo_concurrency.lockutils [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Acquiring lock "9a6720ef-5307-44b7-b566-8d553c9b5384-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.552471] env[63021]: DEBUG oslo_concurrency.lockutils [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Lock "9a6720ef-5307-44b7-b566-8d553c9b5384-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.552688] env[63021]: DEBUG oslo_concurrency.lockutils [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Lock "9a6720ef-5307-44b7-b566-8d553c9b5384-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.555344] env[63021]: INFO nova.compute.manager [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Terminating instance [ 830.560231] env[63021]: DEBUG oslo_concurrency.lockutils [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Acquiring lock "refresh_cache-9a6720ef-5307-44b7-b566-8d553c9b5384" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.560414] env[63021]: DEBUG oslo_concurrency.lockutils [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Acquired lock "refresh_cache-9a6720ef-5307-44b7-b566-8d553c9b5384" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.560661] env[63021]: DEBUG nova.network.neutron [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 830.976711] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': task-1293686, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.029301] env[63021]: DEBUG oslo_concurrency.lockutils [None req-33d0c6a0-b1e3-4677-8b11-5524ccbcf1cd tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Lock "df0b6974-7f78-4b73-8583-d09754fbacea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 96.055s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.035715] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 831.035715] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=63021) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 831.035715] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=63021) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 831.040314] env[63021]: INFO nova.compute.manager [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Took 30.74 seconds to build instance. [ 831.083142] env[63021]: DEBUG nova.network.neutron [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 831.145173] env[63021]: DEBUG nova.network.neutron [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.361543] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf0d6282-f5ad-4918-be35-64009d62f6a6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.372372] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a17f5e0-cd76-4699-af2d-7fc0fe1b735a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.403144] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ced9c5aa-7753-4c26-a806-0b95601d8259 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.410516] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bdc71f6-24a2-47ee-b2a7-81ec48c0b0d4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.423853] env[63021]: DEBUG nova.compute.provider_tree [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 831.476212] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': task-1293686, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.62031} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.476473] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 36d26354-6127-4557-acb0-8743e27ff1ff/36d26354-6127-4557-acb0-8743e27ff1ff.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 831.476687] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 831.476924] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-499c919c-d8ae-4198-b463-bdb3167252cb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.482796] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Waiting for the task: (returnval){ [ 831.482796] env[63021]: value = "task-1293688" [ 831.482796] env[63021]: _type = "Task" [ 831.482796] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.490974] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': task-1293688, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.495493] env[63021]: INFO nova.compute.manager [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Rescuing [ 831.495720] env[63021]: DEBUG oslo_concurrency.lockutils [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquiring lock "refresh_cache-df0b6974-7f78-4b73-8583-d09754fbacea" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.495875] env[63021]: DEBUG oslo_concurrency.lockutils [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquired lock "refresh_cache-df0b6974-7f78-4b73-8583-d09754fbacea" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.496060] env[63021]: DEBUG nova.network.neutron [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 831.542664] env[63021]: DEBUG nova.compute.manager [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 831.546705] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lock "2209d709-7a5b-42e2-be93-b4fd436b6290" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 91.964s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.647889] env[63021]: DEBUG oslo_concurrency.lockutils [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Releasing lock "refresh_cache-9a6720ef-5307-44b7-b566-8d553c9b5384" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.648363] env[63021]: DEBUG nova.compute.manager [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 831.648560] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 831.649447] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6adbccb4-50a8-48ff-b182-80ce4a1f565d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.658945] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 831.659209] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1c64addd-b2c0-4d7b-a09a-6e53f0c80696 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.665023] env[63021]: DEBUG oslo_vmware.api [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Waiting for the task: (returnval){ [ 831.665023] env[63021]: value = "task-1293689" [ 831.665023] env[63021]: _type = "Task" [ 831.665023] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.672470] env[63021]: DEBUG oslo_vmware.api [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293689, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.928313] env[63021]: DEBUG nova.scheduler.client.report [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 831.993225] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': task-1293688, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.097842} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.993581] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 831.994496] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11dfb40a-84c8-4542-b328-a88af81ea9a2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.020011] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] 36d26354-6127-4557-acb0-8743e27ff1ff/36d26354-6127-4557-acb0-8743e27ff1ff.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 832.021342] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-73ca411b-6dcc-4b9a-962a-b40ac68cfd66 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.042063] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Waiting for the task: (returnval){ [ 832.042063] env[63021]: value = "task-1293690" [ 832.042063] env[63021]: _type = "Task" [ 832.042063] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.052063] env[63021]: DEBUG nova.compute.manager [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 832.054423] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': task-1293690, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.071280] env[63021]: DEBUG oslo_concurrency.lockutils [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.174540] env[63021]: DEBUG oslo_vmware.api [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293689, 'name': PowerOffVM_Task, 'duration_secs': 0.409359} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.174851] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 832.174992] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 832.175269] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7eca5c12-91e8-40b9-9a3d-c4793b2c74cd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.198989] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 832.199234] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Deleting contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 832.199418] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Deleting the datastore file [datastore2] 9a6720ef-5307-44b7-b566-8d553c9b5384 {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 832.199689] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eb90d36a-f30b-4580-a1cd-c819dbcd5a54 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.206213] env[63021]: DEBUG oslo_vmware.api [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Waiting for the task: (returnval){ [ 832.206213] env[63021]: value = "task-1293692" [ 832.206213] env[63021]: _type = "Task" [ 832.206213] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.214319] env[63021]: DEBUG oslo_vmware.api [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293692, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.401219] env[63021]: DEBUG nova.network.neutron [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Updating instance_info_cache with network_info: [{"id": "948a6d57-d04a-4c5c-ae6f-cd5786ec9a89", "address": "fa:16:3e:b6:53:e3", "network": {"id": "37008b1f-68ff-404c-9afc-14b108af85c3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1473182647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "2d364c9eae4b4f8a8b60d4a3ac2ed648", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "02bbcead-d833-4543-bec6-fb82dfe659ff", "external-id": "nsx-vlan-transportzone-478", "segmentation_id": 478, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap948a6d57-d0", "ovs_interfaceid": "948a6d57-d04a-4c5c-ae6f-cd5786ec9a89", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.432341] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63021) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 832.432587] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 9.519s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.432837] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.327s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.434430] env[63021]: INFO nova.compute.claims [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 832.551864] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': task-1293690, 'name': ReconfigVM_Task, 'duration_secs': 0.400899} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.552252] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Reconfigured VM instance instance-0000003d to attach disk [datastore1] 36d26354-6127-4557-acb0-8743e27ff1ff/36d26354-6127-4557-acb0-8743e27ff1ff.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 832.552893] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-50099488-a7fa-48ff-b40b-daf3a076e130 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.562579] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Waiting for the task: (returnval){ [ 832.562579] env[63021]: value = "task-1293693" [ 832.562579] env[63021]: _type = "Task" [ 832.562579] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.570837] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': task-1293693, 'name': Rename_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.573379] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.716192] env[63021]: DEBUG oslo_vmware.api [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Task: {'id': task-1293692, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.10827} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.716469] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 832.716686] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Deleted contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 832.716886] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 832.717085] env[63021]: INFO nova.compute.manager [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Took 1.07 seconds to destroy the instance on the hypervisor. [ 832.717396] env[63021]: DEBUG oslo.service.loopingcall [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 832.717626] env[63021]: DEBUG nova.compute.manager [-] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 832.717723] env[63021]: DEBUG nova.network.neutron [-] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 832.736608] env[63021]: DEBUG nova.network.neutron [-] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 832.908096] env[63021]: DEBUG oslo_concurrency.lockutils [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Releasing lock "refresh_cache-df0b6974-7f78-4b73-8583-d09754fbacea" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.072418] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': task-1293693, 'name': Rename_Task, 'duration_secs': 0.145529} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.072739] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 833.072937] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0b76c197-9a01-43f9-96a0-6aa845cddb80 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.080087] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Waiting for the task: (returnval){ [ 833.080087] env[63021]: value = "task-1293695" [ 833.080087] env[63021]: _type = "Task" [ 833.080087] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.088746] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': task-1293695, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.240080] env[63021]: DEBUG nova.network.neutron [-] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.439719] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 833.440032] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d82390b1-ceef-4421-b339-d1fadce4aeca {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.450985] env[63021]: DEBUG oslo_vmware.api [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 833.450985] env[63021]: value = "task-1293696" [ 833.450985] env[63021]: _type = "Task" [ 833.450985] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.458914] env[63021]: DEBUG oslo_vmware.api [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293696, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.591660] env[63021]: DEBUG oslo_vmware.api [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': task-1293695, 'name': PowerOnVM_Task, 'duration_secs': 0.44524} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.591952] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 833.592178] env[63021]: INFO nova.compute.manager [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Took 8.65 seconds to spawn the instance on the hypervisor. [ 833.592353] env[63021]: DEBUG nova.compute.manager [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 833.593122] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a78cad40-7c0d-4874-bde2-2d66e807e4be {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.712146] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07f22cf3-5662-45d8-ba75-dd4b7baf4969 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.719534] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df9d9d9c-9f81-4fa5-9065-711a20fd5059 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.749480] env[63021]: INFO nova.compute.manager [-] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Took 1.03 seconds to deallocate network for instance. [ 833.751867] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1251e05-6eda-48e5-b126-cf654c035a63 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.762031] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bb818ee-1468-4b26-a2fd-98e34057c406 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.777064] env[63021]: DEBUG nova.compute.provider_tree [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 833.960967] env[63021]: DEBUG oslo_vmware.api [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293696, 'name': PowerOffVM_Task, 'duration_secs': 0.411539} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.961247] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 833.962020] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5d6b6db-3b91-4163-8c66-eb29eb8058e4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.979641] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85cafbea-a622-4b08-a045-7837a17f4cba {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.007889] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 834.008185] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fbf9ae6a-fea7-4fcd-b284-1d40755bb29e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.014644] env[63021]: DEBUG oslo_vmware.api [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 834.014644] env[63021]: value = "task-1293697" [ 834.014644] env[63021]: _type = "Task" [ 834.014644] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.021984] env[63021]: DEBUG oslo_vmware.api [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293697, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.112191] env[63021]: INFO nova.compute.manager [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Took 31.95 seconds to build instance. [ 834.258732] env[63021]: DEBUG oslo_concurrency.lockutils [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.279993] env[63021]: DEBUG nova.scheduler.client.report [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 834.525277] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] VM already powered off {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 834.525506] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 834.525743] env[63021]: DEBUG oslo_concurrency.lockutils [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.525891] env[63021]: DEBUG oslo_concurrency.lockutils [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.526084] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 834.526366] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2a727e4f-a070-459e-ae4f-15a961c30b7e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.534369] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 834.534545] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 834.535401] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f09d086-a5f2-4168-bd0d-cb70cd06b6a9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.540525] env[63021]: DEBUG oslo_vmware.api [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 834.540525] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]521a5bd3-881d-f834-c72a-a3d27a4732b9" [ 834.540525] env[63021]: _type = "Task" [ 834.540525] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.551014] env[63021]: DEBUG oslo_vmware.api [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]521a5bd3-881d-f834-c72a-a3d27a4732b9, 'name': SearchDatastore_Task, 'duration_secs': 0.007949} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.552025] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8aea325e-28e7-441a-a1a8-ad210545d93c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.557426] env[63021]: DEBUG oslo_vmware.api [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 834.557426] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52b7a2d7-8cc6-5f36-7af5-7bfa0c7c8b3c" [ 834.557426] env[63021]: _type = "Task" [ 834.557426] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.566068] env[63021]: DEBUG oslo_vmware.api [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52b7a2d7-8cc6-5f36-7af5-7bfa0c7c8b3c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.613782] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5725888a-4958-44b6-8225-8e791048f45a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lock "36d26354-6127-4557-acb0-8743e27ff1ff" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 94.995s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.785021] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.352s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.785590] env[63021]: DEBUG nova.compute.manager [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 834.788761] env[63021]: DEBUG oslo_concurrency.lockutils [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.430s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.790298] env[63021]: INFO nova.compute.claims [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 834.844041] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquiring lock "2209d709-7a5b-42e2-be93-b4fd436b6290" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.844041] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lock "2209d709-7a5b-42e2-be93-b4fd436b6290" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.844295] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquiring lock "2209d709-7a5b-42e2-be93-b4fd436b6290-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.844295] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lock "2209d709-7a5b-42e2-be93-b4fd436b6290-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.844482] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lock "2209d709-7a5b-42e2-be93-b4fd436b6290-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.846665] env[63021]: INFO nova.compute.manager [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Terminating instance [ 834.848571] env[63021]: DEBUG nova.compute.manager [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 834.848783] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 834.849707] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4a92a43-40bd-4cc5-acbb-bd579286e082 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.857990] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 834.858243] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6dd7a05c-02c1-4921-9f58-7a5103bc916d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.865101] env[63021]: DEBUG oslo_vmware.api [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Waiting for the task: (returnval){ [ 834.865101] env[63021]: value = "task-1293699" [ 834.865101] env[63021]: _type = "Task" [ 834.865101] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.873807] env[63021]: DEBUG oslo_vmware.api [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': task-1293699, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.915171] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquiring lock "36d26354-6127-4557-acb0-8743e27ff1ff" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.915503] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lock "36d26354-6127-4557-acb0-8743e27ff1ff" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.915728] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquiring lock "36d26354-6127-4557-acb0-8743e27ff1ff-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.915913] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lock "36d26354-6127-4557-acb0-8743e27ff1ff-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.916102] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lock "36d26354-6127-4557-acb0-8743e27ff1ff-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.918618] env[63021]: INFO nova.compute.manager [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Terminating instance [ 834.920699] env[63021]: DEBUG nova.compute.manager [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 834.920699] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 834.921562] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4461c82-2f42-423c-815c-4b62a78fa607 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.930230] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 834.930510] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-915369b1-fba9-4980-b7ba-60ba6d76766b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.937596] env[63021]: DEBUG oslo_vmware.api [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Waiting for the task: (returnval){ [ 834.937596] env[63021]: value = "task-1293700" [ 834.937596] env[63021]: _type = "Task" [ 834.937596] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.946293] env[63021]: DEBUG oslo_vmware.api [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': task-1293700, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.069638] env[63021]: DEBUG oslo_vmware.api [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52b7a2d7-8cc6-5f36-7af5-7bfa0c7c8b3c, 'name': SearchDatastore_Task, 'duration_secs': 0.008335} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.069966] env[63021]: DEBUG oslo_concurrency.lockutils [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.070290] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] df0b6974-7f78-4b73-8583-d09754fbacea/e624225d-af0f-415c-90cc-d4b228c479f6-rescue.vmdk. {{(pid=63021) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 835.070625] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1a146bb7-5a55-4038-b709-51f1c675457c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.077012] env[63021]: DEBUG oslo_vmware.api [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 835.077012] env[63021]: value = "task-1293701" [ 835.077012] env[63021]: _type = "Task" [ 835.077012] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.085107] env[63021]: DEBUG oslo_vmware.api [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293701, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.116365] env[63021]: DEBUG nova.compute.manager [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 835.295028] env[63021]: DEBUG nova.compute.utils [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 835.299591] env[63021]: DEBUG nova.compute.manager [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 835.299591] env[63021]: DEBUG nova.network.neutron [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 835.354589] env[63021]: DEBUG nova.policy [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fb21dfc020df428a931f7500484826ab', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '038a9a0a01e04b32b023d9dc678b30aa', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 835.380198] env[63021]: DEBUG oslo_vmware.api [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': task-1293699, 'name': PowerOffVM_Task, 'duration_secs': 0.197529} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.380420] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 835.380591] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 835.380860] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-418d9677-a9a7-4b4c-9de4-8003c45ad193 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.448598] env[63021]: DEBUG oslo_vmware.api [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': task-1293700, 'name': PowerOffVM_Task, 'duration_secs': 0.236141} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.450158] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 835.450339] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 835.450606] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 835.450815] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Deleting contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 835.451063] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Deleting the datastore file [datastore2] 2209d709-7a5b-42e2-be93-b4fd436b6290 {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 835.451303] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-71bd2a5d-b12d-4de3-8d6f-cfbac6f6bde3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.453463] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c0cbda3c-d277-4e13-baf1-c22cf188e9b4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.461980] env[63021]: DEBUG oslo_vmware.api [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Waiting for the task: (returnval){ [ 835.461980] env[63021]: value = "task-1293704" [ 835.461980] env[63021]: _type = "Task" [ 835.461980] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.471127] env[63021]: DEBUG oslo_vmware.api [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': task-1293704, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.525278] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 835.525535] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 835.525740] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Deleting the datastore file [datastore1] 36d26354-6127-4557-acb0-8743e27ff1ff {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 835.526043] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d734c40d-45b0-4acb-840c-ff9628d6d46d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.532827] env[63021]: DEBUG oslo_vmware.api [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Waiting for the task: (returnval){ [ 835.532827] env[63021]: value = "task-1293705" [ 835.532827] env[63021]: _type = "Task" [ 835.532827] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.543375] env[63021]: DEBUG oslo_vmware.api [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': task-1293705, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.588914] env[63021]: DEBUG oslo_vmware.api [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293701, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.509463} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.589321] env[63021]: INFO nova.virt.vmwareapi.ds_util [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] df0b6974-7f78-4b73-8583-d09754fbacea/e624225d-af0f-415c-90cc-d4b228c479f6-rescue.vmdk. [ 835.590273] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85f0d279-ce1a-424c-aca2-c81ed246196a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.619307] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] df0b6974-7f78-4b73-8583-d09754fbacea/e624225d-af0f-415c-90cc-d4b228c479f6-rescue.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 835.619307] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c0fd186e-c33e-41c1-9d6c-c57b4bc0d222 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.641617] env[63021]: DEBUG oslo_vmware.api [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 835.641617] env[63021]: value = "task-1293706" [ 835.641617] env[63021]: _type = "Task" [ 835.641617] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.646292] env[63021]: DEBUG nova.network.neutron [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Successfully created port: d82f4ff7-a0f2-401a-8b11-1fdcc8ba0bc8 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 835.654106] env[63021]: DEBUG oslo_vmware.api [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293706, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.660026] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.800116] env[63021]: DEBUG nova.compute.manager [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 835.971979] env[63021]: DEBUG oslo_vmware.api [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': task-1293704, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.245554} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.974784] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 835.975075] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Deleted contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 835.975337] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 835.975547] env[63021]: INFO nova.compute.manager [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Took 1.13 seconds to destroy the instance on the hypervisor. [ 835.975882] env[63021]: DEBUG oslo.service.loopingcall [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 835.976328] env[63021]: DEBUG nova.compute.manager [-] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 835.976928] env[63021]: DEBUG nova.network.neutron [-] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 836.042833] env[63021]: DEBUG oslo_vmware.api [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Task: {'id': task-1293705, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.191656} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.046190] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 836.046190] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 836.046190] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 836.046190] env[63021]: INFO nova.compute.manager [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Took 1.13 seconds to destroy the instance on the hypervisor. [ 836.046415] env[63021]: DEBUG oslo.service.loopingcall [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 836.046783] env[63021]: DEBUG nova.compute.manager [-] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 836.046924] env[63021]: DEBUG nova.network.neutron [-] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 836.130121] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15279d3a-e5df-4f8b-b33f-28cbffd8b88c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.138561] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67afdcf3-b0e9-427f-9be1-41ce92040ca0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.182672] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea2ae82b-da49-41f6-8066-98d8e68d3032 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.190222] env[63021]: DEBUG oslo_vmware.api [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293706, 'name': ReconfigVM_Task, 'duration_secs': 0.293253} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.190614] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Reconfigured VM instance instance-0000003b to attach disk [datastore2] df0b6974-7f78-4b73-8583-d09754fbacea/e624225d-af0f-415c-90cc-d4b228c479f6-rescue.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 836.193132] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7eb9773-eb9e-497d-891f-aca864d70018 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.198548] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceecdcc0-66a4-4954-b5ca-f96a7ef3f835 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.226877] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-47e2d6fe-9311-4daf-a93d-cb13bac69304 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.246159] env[63021]: DEBUG nova.compute.provider_tree [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 836.249972] env[63021]: DEBUG nova.compute.manager [req-19b3997e-098e-41b3-9364-9f8cdefc7e13 req-5a9c6046-4be5-4bab-8200-153e4803703e service nova] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Received event network-vif-deleted-ad4b54c8-9cd8-4dde-b661-29b32a858e71 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 836.250194] env[63021]: INFO nova.compute.manager [req-19b3997e-098e-41b3-9364-9f8cdefc7e13 req-5a9c6046-4be5-4bab-8200-153e4803703e service nova] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Neutron deleted interface ad4b54c8-9cd8-4dde-b661-29b32a858e71; detaching it from the instance and deleting it from the info cache [ 836.250367] env[63021]: DEBUG nova.network.neutron [req-19b3997e-098e-41b3-9364-9f8cdefc7e13 req-5a9c6046-4be5-4bab-8200-153e4803703e service nova] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.257724] env[63021]: DEBUG oslo_vmware.api [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 836.257724] env[63021]: value = "task-1293707" [ 836.257724] env[63021]: _type = "Task" [ 836.257724] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.266532] env[63021]: DEBUG oslo_vmware.api [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293707, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.365039] env[63021]: DEBUG nova.compute.manager [req-71ec487d-df46-4ae5-8be4-9a1bae1a82fd req-e64067ac-8412-466e-a73a-9279cbab8588 service nova] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Received event network-vif-deleted-8fee9441-e591-4a58-8296-3ec48ab9c154 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 836.365039] env[63021]: INFO nova.compute.manager [req-71ec487d-df46-4ae5-8be4-9a1bae1a82fd req-e64067ac-8412-466e-a73a-9279cbab8588 service nova] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Neutron deleted interface 8fee9441-e591-4a58-8296-3ec48ab9c154; detaching it from the instance and deleting it from the info cache [ 836.365139] env[63021]: DEBUG nova.network.neutron [req-71ec487d-df46-4ae5-8be4-9a1bae1a82fd req-e64067ac-8412-466e-a73a-9279cbab8588 service nova] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.746725] env[63021]: DEBUG nova.network.neutron [-] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.752532] env[63021]: DEBUG nova.scheduler.client.report [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 836.758145] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0f5d78c5-d340-41bd-b99d-557158dc76c3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.773713] env[63021]: DEBUG oslo_vmware.api [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293707, 'name': ReconfigVM_Task, 'duration_secs': 0.199128} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.775289] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 836.775981] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f2673031-097c-4e29-a455-b1d508593072 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.785127] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d6697fa-7a9d-446e-a7de-bf1105119d62 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.800112] env[63021]: DEBUG nova.network.neutron [-] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.806747] env[63021]: DEBUG oslo_vmware.api [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 836.806747] env[63021]: value = "task-1293708" [ 836.806747] env[63021]: _type = "Task" [ 836.806747] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.816268] env[63021]: DEBUG nova.compute.manager [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 836.818385] env[63021]: DEBUG nova.compute.manager [req-19b3997e-098e-41b3-9364-9f8cdefc7e13 req-5a9c6046-4be5-4bab-8200-153e4803703e service nova] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Detach interface failed, port_id=ad4b54c8-9cd8-4dde-b661-29b32a858e71, reason: Instance 2209d709-7a5b-42e2-be93-b4fd436b6290 could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 836.823781] env[63021]: DEBUG oslo_vmware.api [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293708, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.844353] env[63021]: DEBUG nova.virt.hardware [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 836.844868] env[63021]: DEBUG nova.virt.hardware [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 836.844868] env[63021]: DEBUG nova.virt.hardware [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 836.845028] env[63021]: DEBUG nova.virt.hardware [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 836.845443] env[63021]: DEBUG nova.virt.hardware [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 836.845443] env[63021]: DEBUG nova.virt.hardware [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 836.845614] env[63021]: DEBUG nova.virt.hardware [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 836.845802] env[63021]: DEBUG nova.virt.hardware [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 836.845990] env[63021]: DEBUG nova.virt.hardware [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 836.846198] env[63021]: DEBUG nova.virt.hardware [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 836.846379] env[63021]: DEBUG nova.virt.hardware [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 836.847241] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b25e531-1211-4755-a1db-581fb39e3250 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.855476] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f2fb2d5-1d88-4345-8dc8-4684a05e7e4d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.871476] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b73f1cb5-ee33-42f0-a48d-0aaadcff57ba {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.879674] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-603ce8a4-e295-4e71-8efc-ec2826062ab9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.904422] env[63021]: DEBUG nova.compute.manager [req-71ec487d-df46-4ae5-8be4-9a1bae1a82fd req-e64067ac-8412-466e-a73a-9279cbab8588 service nova] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Detach interface failed, port_id=8fee9441-e591-4a58-8296-3ec48ab9c154, reason: Instance 36d26354-6127-4557-acb0-8743e27ff1ff could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 837.174555] env[63021]: DEBUG nova.network.neutron [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Successfully updated port: d82f4ff7-a0f2-401a-8b11-1fdcc8ba0bc8 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 837.249561] env[63021]: INFO nova.compute.manager [-] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Took 1.27 seconds to deallocate network for instance. [ 837.258885] env[63021]: DEBUG oslo_concurrency.lockutils [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.470s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.259696] env[63021]: DEBUG nova.compute.manager [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 837.262540] env[63021]: DEBUG oslo_concurrency.lockutils [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 18.427s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.262876] env[63021]: DEBUG nova.objects.instance [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63021) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 837.301738] env[63021]: INFO nova.compute.manager [-] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Took 1.25 seconds to deallocate network for instance. [ 837.318818] env[63021]: DEBUG oslo_vmware.api [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293708, 'name': PowerOnVM_Task, 'duration_secs': 0.398656} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.319281] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 837.322431] env[63021]: DEBUG nova.compute.manager [None req-90365dbc-355f-4a8a-9258-8ea32f910a5a tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 837.323501] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3736cf4-1eba-44a7-a854-4c192f7b7a7e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.677762] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Acquiring lock "refresh_cache-2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.677926] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Acquired lock "refresh_cache-2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.678097] env[63021]: DEBUG nova.network.neutron [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 837.757703] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.775103] env[63021]: DEBUG nova.compute.utils [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 837.776508] env[63021]: DEBUG nova.compute.manager [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Not allocating networking since 'none' was specified. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 837.813523] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.212391] env[63021]: DEBUG nova.network.neutron [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 838.278238] env[63021]: DEBUG oslo_concurrency.lockutils [None req-777a4a6a-1ba0-45f1-9234-bfd5ce14d3cc tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.281737] env[63021]: DEBUG oslo_concurrency.lockutils [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.120s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.283306] env[63021]: INFO nova.compute.claims [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 838.286761] env[63021]: DEBUG nova.compute.manager [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 838.366670] env[63021]: DEBUG nova.network.neutron [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Updating instance_info_cache with network_info: [{"id": "d82f4ff7-a0f2-401a-8b11-1fdcc8ba0bc8", "address": "fa:16:3e:66:5b:e3", "network": {"id": "f8ef73a4-dfd1-4434-8490-19f156d1e01a", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1357446952-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "038a9a0a01e04b32b023d9dc678b30aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "195e328b-e41a-49f5-9e51-546b8ea8ceba", "external-id": "nsx-vlan-transportzone-735", "segmentation_id": 735, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd82f4ff7-a0", "ovs_interfaceid": "d82f4ff7-a0f2-401a-8b11-1fdcc8ba0bc8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.515076] env[63021]: DEBUG nova.compute.manager [req-50f71d74-4a75-4c65-b4a3-a680fe0139fa req-6f3587cc-7ea7-4a13-a0eb-03af0ebb229b service nova] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Received event network-vif-plugged-d82f4ff7-a0f2-401a-8b11-1fdcc8ba0bc8 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 838.515370] env[63021]: DEBUG oslo_concurrency.lockutils [req-50f71d74-4a75-4c65-b4a3-a680fe0139fa req-6f3587cc-7ea7-4a13-a0eb-03af0ebb229b service nova] Acquiring lock "2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.515560] env[63021]: DEBUG oslo_concurrency.lockutils [req-50f71d74-4a75-4c65-b4a3-a680fe0139fa req-6f3587cc-7ea7-4a13-a0eb-03af0ebb229b service nova] Lock "2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.515744] env[63021]: DEBUG oslo_concurrency.lockutils [req-50f71d74-4a75-4c65-b4a3-a680fe0139fa req-6f3587cc-7ea7-4a13-a0eb-03af0ebb229b service nova] Lock "2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.515944] env[63021]: DEBUG nova.compute.manager [req-50f71d74-4a75-4c65-b4a3-a680fe0139fa req-6f3587cc-7ea7-4a13-a0eb-03af0ebb229b service nova] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] No waiting events found dispatching network-vif-plugged-d82f4ff7-a0f2-401a-8b11-1fdcc8ba0bc8 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 838.516362] env[63021]: WARNING nova.compute.manager [req-50f71d74-4a75-4c65-b4a3-a680fe0139fa req-6f3587cc-7ea7-4a13-a0eb-03af0ebb229b service nova] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Received unexpected event network-vif-plugged-d82f4ff7-a0f2-401a-8b11-1fdcc8ba0bc8 for instance with vm_state building and task_state spawning. [ 838.516542] env[63021]: DEBUG nova.compute.manager [req-50f71d74-4a75-4c65-b4a3-a680fe0139fa req-6f3587cc-7ea7-4a13-a0eb-03af0ebb229b service nova] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Received event network-changed-d82f4ff7-a0f2-401a-8b11-1fdcc8ba0bc8 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 838.516701] env[63021]: DEBUG nova.compute.manager [req-50f71d74-4a75-4c65-b4a3-a680fe0139fa req-6f3587cc-7ea7-4a13-a0eb-03af0ebb229b service nova] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Refreshing instance network info cache due to event network-changed-d82f4ff7-a0f2-401a-8b11-1fdcc8ba0bc8. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 838.516872] env[63021]: DEBUG oslo_concurrency.lockutils [req-50f71d74-4a75-4c65-b4a3-a680fe0139fa req-6f3587cc-7ea7-4a13-a0eb-03af0ebb229b service nova] Acquiring lock "refresh_cache-2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.870048] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Releasing lock "refresh_cache-2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.870048] env[63021]: DEBUG nova.compute.manager [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Instance network_info: |[{"id": "d82f4ff7-a0f2-401a-8b11-1fdcc8ba0bc8", "address": "fa:16:3e:66:5b:e3", "network": {"id": "f8ef73a4-dfd1-4434-8490-19f156d1e01a", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1357446952-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "038a9a0a01e04b32b023d9dc678b30aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "195e328b-e41a-49f5-9e51-546b8ea8ceba", "external-id": "nsx-vlan-transportzone-735", "segmentation_id": 735, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd82f4ff7-a0", "ovs_interfaceid": "d82f4ff7-a0f2-401a-8b11-1fdcc8ba0bc8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 838.870256] env[63021]: DEBUG oslo_concurrency.lockutils [req-50f71d74-4a75-4c65-b4a3-a680fe0139fa req-6f3587cc-7ea7-4a13-a0eb-03af0ebb229b service nova] Acquired lock "refresh_cache-2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.870363] env[63021]: DEBUG nova.network.neutron [req-50f71d74-4a75-4c65-b4a3-a680fe0139fa req-6f3587cc-7ea7-4a13-a0eb-03af0ebb229b service nova] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Refreshing network info cache for port d82f4ff7-a0f2-401a-8b11-1fdcc8ba0bc8 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 838.871654] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:66:5b:e3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '195e328b-e41a-49f5-9e51-546b8ea8ceba', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd82f4ff7-a0f2-401a-8b11-1fdcc8ba0bc8', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 838.881337] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Creating folder: Project (038a9a0a01e04b32b023d9dc678b30aa). Parent ref: group-v277447. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 838.882335] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-53c3f8f7-4073-4adb-9be6-1d9ae4c3a6f2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.893447] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Created folder: Project (038a9a0a01e04b32b023d9dc678b30aa) in parent group-v277447. [ 838.893636] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Creating folder: Instances. Parent ref: group-v277494. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 838.893903] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-74ff51d9-a6a7-465f-8e36-ad4f8c90cd81 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.902468] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Created folder: Instances in parent group-v277494. [ 838.902695] env[63021]: DEBUG oslo.service.loopingcall [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 838.902887] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 838.903235] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-249baaba-8f8a-4983-8305-539354402584 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.922925] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 838.922925] env[63021]: value = "task-1293713" [ 838.922925] env[63021]: _type = "Task" [ 838.922925] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.930323] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293713, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.144424] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquiring lock "1b4871cc-3dac-4683-8fce-93c4d9e4407d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.144655] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Lock "1b4871cc-3dac-4683-8fce-93c4d9e4407d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.300413] env[63021]: DEBUG nova.compute.manager [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 839.327197] env[63021]: DEBUG nova.virt.hardware [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 839.327464] env[63021]: DEBUG nova.virt.hardware [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 839.327619] env[63021]: DEBUG nova.virt.hardware [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 839.327796] env[63021]: DEBUG nova.virt.hardware [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 839.327939] env[63021]: DEBUG nova.virt.hardware [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 839.328098] env[63021]: DEBUG nova.virt.hardware [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 839.328305] env[63021]: DEBUG nova.virt.hardware [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 839.328459] env[63021]: DEBUG nova.virt.hardware [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 839.328624] env[63021]: DEBUG nova.virt.hardware [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 839.328786] env[63021]: DEBUG nova.virt.hardware [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 839.328954] env[63021]: DEBUG nova.virt.hardware [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 839.329817] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cfef572-a975-4722-be48-48fe1a79dc65 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.340118] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0068221c-f2bc-48cf-a0f1-0e4e757c1cd9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.355346] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Instance VIF info [] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 839.360682] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Creating folder: Project (c8021749fc1d4146a432758bcfa0ba2c). Parent ref: group-v277447. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 839.360968] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a9c90255-7a1c-49de-8803-6a471510c43f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.370344] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Created folder: Project (c8021749fc1d4146a432758bcfa0ba2c) in parent group-v277447. [ 839.370533] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Creating folder: Instances. Parent ref: group-v277497. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 839.370758] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-93b22e27-2c78-412b-8c02-da8be5d66407 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.379590] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Created folder: Instances in parent group-v277497. [ 839.379816] env[63021]: DEBUG oslo.service.loopingcall [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 839.380007] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 839.380217] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-47cabbe4-b112-4b77-b987-69e11338da96 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.400783] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 839.400783] env[63021]: value = "task-1293716" [ 839.400783] env[63021]: _type = "Task" [ 839.400783] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.409961] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293716, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.433717] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293713, 'name': CreateVM_Task, 'duration_secs': 0.367333} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.433813] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 839.434743] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.434743] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.434967] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 839.435244] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4545ac76-9e6a-4c0f-9ef1-a0eda729c2d9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.439452] env[63021]: DEBUG oslo_vmware.api [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Waiting for the task: (returnval){ [ 839.439452] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52e47ae4-9602-2299-f514-1eb1037255a8" [ 839.439452] env[63021]: _type = "Task" [ 839.439452] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.449862] env[63021]: DEBUG oslo_vmware.api [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52e47ae4-9602-2299-f514-1eb1037255a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.634430] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c36a69d-424e-47b7-8eab-b8e155097378 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.641749] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42d518ea-dfdd-4906-990f-1f9969ddb155 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.672290] env[63021]: DEBUG nova.network.neutron [req-50f71d74-4a75-4c65-b4a3-a680fe0139fa req-6f3587cc-7ea7-4a13-a0eb-03af0ebb229b service nova] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Updated VIF entry in instance network info cache for port d82f4ff7-a0f2-401a-8b11-1fdcc8ba0bc8. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 839.672648] env[63021]: DEBUG nova.network.neutron [req-50f71d74-4a75-4c65-b4a3-a680fe0139fa req-6f3587cc-7ea7-4a13-a0eb-03af0ebb229b service nova] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Updating instance_info_cache with network_info: [{"id": "d82f4ff7-a0f2-401a-8b11-1fdcc8ba0bc8", "address": "fa:16:3e:66:5b:e3", "network": {"id": "f8ef73a4-dfd1-4434-8490-19f156d1e01a", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1357446952-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "038a9a0a01e04b32b023d9dc678b30aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "195e328b-e41a-49f5-9e51-546b8ea8ceba", "external-id": "nsx-vlan-transportzone-735", "segmentation_id": 735, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd82f4ff7-a0", "ovs_interfaceid": "d82f4ff7-a0f2-401a-8b11-1fdcc8ba0bc8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.674337] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b59a3215-cb1d-44ff-81cd-b9f1adc3bbe5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.682027] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4c3f1fd-c627-4747-b032-545fd26d6e26 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.696273] env[63021]: DEBUG nova.compute.provider_tree [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 839.912632] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293716, 'name': CreateVM_Task, 'duration_secs': 0.264739} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.912820] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 839.913245] env[63021]: DEBUG oslo_concurrency.lockutils [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.952071] env[63021]: DEBUG oslo_vmware.api [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52e47ae4-9602-2299-f514-1eb1037255a8, 'name': SearchDatastore_Task, 'duration_secs': 0.038437} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.952442] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.952681] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 839.952920] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.953104] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.953297] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 839.953554] env[63021]: DEBUG oslo_concurrency.lockutils [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.953852] env[63021]: DEBUG oslo_concurrency.lockutils [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 839.954109] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-04228f02-2997-422f-867e-e9948b2cf7f8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.955964] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e23f45f-9c4e-4f40-ab07-04c476a0978c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.960806] env[63021]: DEBUG oslo_vmware.api [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Waiting for the task: (returnval){ [ 839.960806] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52924eaa-ebee-6b32-7da0-793b8fd6bbe2" [ 839.960806] env[63021]: _type = "Task" [ 839.960806] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.965980] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 839.966248] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 839.967488] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a8c542ba-d9cd-4955-8208-459fef30c644 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.973168] env[63021]: DEBUG oslo_vmware.api [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52924eaa-ebee-6b32-7da0-793b8fd6bbe2, 'name': SearchDatastore_Task, 'duration_secs': 0.009659} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.973764] env[63021]: DEBUG oslo_concurrency.lockutils [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.974023] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 839.974242] env[63021]: DEBUG oslo_concurrency.lockutils [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.975836] env[63021]: DEBUG oslo_vmware.api [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Waiting for the task: (returnval){ [ 839.975836] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]523f42c3-507c-4131-27eb-f064645fc8d8" [ 839.975836] env[63021]: _type = "Task" [ 839.975836] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.983262] env[63021]: DEBUG oslo_vmware.api [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]523f42c3-507c-4131-27eb-f064645fc8d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.178343] env[63021]: DEBUG oslo_concurrency.lockutils [req-50f71d74-4a75-4c65-b4a3-a680fe0139fa req-6f3587cc-7ea7-4a13-a0eb-03af0ebb229b service nova] Releasing lock "refresh_cache-2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.199430] env[63021]: DEBUG nova.scheduler.client.report [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 840.486608] env[63021]: DEBUG oslo_vmware.api [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]523f42c3-507c-4131-27eb-f064645fc8d8, 'name': SearchDatastore_Task, 'duration_secs': 0.008543} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.487421] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4c2b979-fe2b-4075-ab28-6cb41b1a37ea {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.492415] env[63021]: DEBUG oslo_vmware.api [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Waiting for the task: (returnval){ [ 840.492415] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5277bb3b-860f-6c43-aa3d-e43049ee1f6b" [ 840.492415] env[63021]: _type = "Task" [ 840.492415] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.499561] env[63021]: DEBUG oslo_vmware.api [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5277bb3b-860f-6c43-aa3d-e43049ee1f6b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.704709] env[63021]: DEBUG oslo_concurrency.lockutils [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.423s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.705321] env[63021]: DEBUG nova.compute.manager [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 840.708078] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.514s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.709516] env[63021]: INFO nova.compute.claims [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 841.003475] env[63021]: DEBUG oslo_vmware.api [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5277bb3b-860f-6c43-aa3d-e43049ee1f6b, 'name': SearchDatastore_Task, 'duration_secs': 0.008615} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.003774] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.004086] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4/2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 841.004387] env[63021]: DEBUG oslo_concurrency.lockutils [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.004588] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 841.004807] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1f8e0581-89d6-4854-9dac-7e31b64a316c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.006692] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-aa871e60-3183-4e7b-9ef1-c5ade9f5734e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.013427] env[63021]: DEBUG oslo_vmware.api [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Waiting for the task: (returnval){ [ 841.013427] env[63021]: value = "task-1293718" [ 841.013427] env[63021]: _type = "Task" [ 841.013427] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.014541] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 841.014725] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 841.018402] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67b180ff-7191-4e6a-931f-f6a4d6ba90a1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.025767] env[63021]: DEBUG oslo_vmware.api [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Task: {'id': task-1293718, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.027223] env[63021]: DEBUG oslo_vmware.api [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Waiting for the task: (returnval){ [ 841.027223] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]528cefa0-3064-4858-9c30-41675f50b163" [ 841.027223] env[63021]: _type = "Task" [ 841.027223] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.035419] env[63021]: DEBUG oslo_vmware.api [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]528cefa0-3064-4858-9c30-41675f50b163, 'name': SearchDatastore_Task, 'duration_secs': 0.007705} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.036143] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0f3eaf7-6a0b-4a4e-afa7-bb79fcd8f0bd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.040962] env[63021]: DEBUG oslo_vmware.api [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Waiting for the task: (returnval){ [ 841.040962] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]524ff1a3-342d-4f3d-159b-911da356fa32" [ 841.040962] env[63021]: _type = "Task" [ 841.040962] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.048205] env[63021]: DEBUG oslo_vmware.api [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]524ff1a3-342d-4f3d-159b-911da356fa32, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.214618] env[63021]: DEBUG nova.compute.utils [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 841.216161] env[63021]: DEBUG nova.compute.manager [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 841.216368] env[63021]: DEBUG nova.network.neutron [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 841.268034] env[63021]: DEBUG nova.policy [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '196d454130514a9887dcec2adb31dfa6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '70e912b7ac9144c793323eaf26287ba1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 841.523822] env[63021]: DEBUG oslo_vmware.api [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Task: {'id': task-1293718, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.478725} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.524121] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4/2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 841.524343] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 841.524588] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f5578a73-2f72-4673-bff5-8231515cb2ef {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.526934] env[63021]: DEBUG nova.network.neutron [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Successfully created port: e581fd41-04b7-4081-a7be-c6389f4d623f {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 841.532749] env[63021]: DEBUG oslo_vmware.api [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Waiting for the task: (returnval){ [ 841.532749] env[63021]: value = "task-1293719" [ 841.532749] env[63021]: _type = "Task" [ 841.532749] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.540557] env[63021]: DEBUG oslo_vmware.api [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Task: {'id': task-1293719, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.549372] env[63021]: DEBUG oslo_vmware.api [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]524ff1a3-342d-4f3d-159b-911da356fa32, 'name': SearchDatastore_Task, 'duration_secs': 0.007928} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.549594] env[63021]: DEBUG oslo_concurrency.lockutils [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.549847] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 6f4e6245-c0d9-46b4-baf5-5aca19e66da5/6f4e6245-c0d9-46b4-baf5-5aca19e66da5.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 841.550106] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d2592b92-7616-4ff2-b232-1553263fc43b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.555639] env[63021]: DEBUG oslo_vmware.api [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Waiting for the task: (returnval){ [ 841.555639] env[63021]: value = "task-1293720" [ 841.555639] env[63021]: _type = "Task" [ 841.555639] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.562527] env[63021]: DEBUG oslo_vmware.api [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293720, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.721601] env[63021]: DEBUG nova.compute.manager [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 842.043726] env[63021]: DEBUG oslo_vmware.api [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Task: {'id': task-1293719, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074419} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.044064] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 842.046933] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d62974ba-6b80-41a7-bb7f-585265ad5c97 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.068545] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Reconfiguring VM instance instance-0000003e to attach disk [datastore1] 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4/2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 842.073633] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec746946-6a6e-43eb-b2dd-f1977e7d8d1a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.094313] env[63021]: DEBUG oslo_vmware.api [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293720, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.095598] env[63021]: DEBUG oslo_vmware.api [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Waiting for the task: (returnval){ [ 842.095598] env[63021]: value = "task-1293721" [ 842.095598] env[63021]: _type = "Task" [ 842.095598] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.099662] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-723812e3-f071-4638-8956-5da36307ae1a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.106934] env[63021]: DEBUG oslo_vmware.api [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Task: {'id': task-1293721, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.109292] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71bb5e5a-8a95-48cd-a86b-9c54df4aadaa {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.139625] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8f38e78-4833-49ff-90e0-14c0b28884ca {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.146768] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-866437f1-d26e-4a24-9220-675a6663dd1e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.159805] env[63021]: DEBUG nova.compute.provider_tree [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 842.569991] env[63021]: DEBUG oslo_vmware.api [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293720, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.607030] env[63021]: DEBUG oslo_vmware.api [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Task: {'id': task-1293721, 'name': ReconfigVM_Task, 'duration_secs': 0.260618} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.607030] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Reconfigured VM instance instance-0000003e to attach disk [datastore1] 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4/2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 842.607264] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bef3e563-b2c5-4d71-a392-14c8879c5d55 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.613892] env[63021]: DEBUG oslo_vmware.api [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Waiting for the task: (returnval){ [ 842.613892] env[63021]: value = "task-1293722" [ 842.613892] env[63021]: _type = "Task" [ 842.613892] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.621713] env[63021]: DEBUG oslo_vmware.api [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Task: {'id': task-1293722, 'name': Rename_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.663014] env[63021]: DEBUG nova.scheduler.client.report [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 842.733771] env[63021]: DEBUG nova.compute.manager [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 842.759457] env[63021]: DEBUG nova.virt.hardware [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 842.759758] env[63021]: DEBUG nova.virt.hardware [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 842.759919] env[63021]: DEBUG nova.virt.hardware [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 842.760141] env[63021]: DEBUG nova.virt.hardware [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 842.760288] env[63021]: DEBUG nova.virt.hardware [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 842.760433] env[63021]: DEBUG nova.virt.hardware [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 842.760638] env[63021]: DEBUG nova.virt.hardware [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 842.760826] env[63021]: DEBUG nova.virt.hardware [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 842.761000] env[63021]: DEBUG nova.virt.hardware [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 842.761181] env[63021]: DEBUG nova.virt.hardware [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 842.761354] env[63021]: DEBUG nova.virt.hardware [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 842.762322] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4a5d7fe-4172-4f7c-af68-061115557d80 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.770020] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f445017-e7f6-4726-b0b5-0e61193c6a48 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.026841] env[63021]: DEBUG nova.compute.manager [req-b19d5172-c602-473b-bb7b-885dc0761a6a req-3bda21ce-ba38-4ac8-96ea-85eaffcf31a2 service nova] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Received event network-vif-plugged-e581fd41-04b7-4081-a7be-c6389f4d623f {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 843.027149] env[63021]: DEBUG oslo_concurrency.lockutils [req-b19d5172-c602-473b-bb7b-885dc0761a6a req-3bda21ce-ba38-4ac8-96ea-85eaffcf31a2 service nova] Acquiring lock "7787f6bb-e855-476f-9146-2a648932429d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.027307] env[63021]: DEBUG oslo_concurrency.lockutils [req-b19d5172-c602-473b-bb7b-885dc0761a6a req-3bda21ce-ba38-4ac8-96ea-85eaffcf31a2 service nova] Lock "7787f6bb-e855-476f-9146-2a648932429d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.027451] env[63021]: DEBUG oslo_concurrency.lockutils [req-b19d5172-c602-473b-bb7b-885dc0761a6a req-3bda21ce-ba38-4ac8-96ea-85eaffcf31a2 service nova] Lock "7787f6bb-e855-476f-9146-2a648932429d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.027614] env[63021]: DEBUG nova.compute.manager [req-b19d5172-c602-473b-bb7b-885dc0761a6a req-3bda21ce-ba38-4ac8-96ea-85eaffcf31a2 service nova] [instance: 7787f6bb-e855-476f-9146-2a648932429d] No waiting events found dispatching network-vif-plugged-e581fd41-04b7-4081-a7be-c6389f4d623f {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 843.027780] env[63021]: WARNING nova.compute.manager [req-b19d5172-c602-473b-bb7b-885dc0761a6a req-3bda21ce-ba38-4ac8-96ea-85eaffcf31a2 service nova] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Received unexpected event network-vif-plugged-e581fd41-04b7-4081-a7be-c6389f4d623f for instance with vm_state building and task_state spawning. [ 843.071131] env[63021]: DEBUG oslo_vmware.api [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293720, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.476378} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.071338] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 6f4e6245-c0d9-46b4-baf5-5aca19e66da5/6f4e6245-c0d9-46b4-baf5-5aca19e66da5.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 843.071519] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 843.071758] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5f76c124-98fe-4574-bf6d-803b9dd24946 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.078723] env[63021]: DEBUG oslo_vmware.api [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Waiting for the task: (returnval){ [ 843.078723] env[63021]: value = "task-1293724" [ 843.078723] env[63021]: _type = "Task" [ 843.078723] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.087106] env[63021]: DEBUG oslo_vmware.api [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293724, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.115022] env[63021]: DEBUG nova.network.neutron [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Successfully updated port: e581fd41-04b7-4081-a7be-c6389f4d623f {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 843.125780] env[63021]: DEBUG oslo_vmware.api [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Task: {'id': task-1293722, 'name': Rename_Task, 'duration_secs': 0.177354} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.126077] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 843.126329] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-74822c96-5cf9-4e4e-b79e-31846b952097 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.133731] env[63021]: DEBUG oslo_vmware.api [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Waiting for the task: (returnval){ [ 843.133731] env[63021]: value = "task-1293725" [ 843.133731] env[63021]: _type = "Task" [ 843.133731] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.145172] env[63021]: DEBUG oslo_vmware.api [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Task: {'id': task-1293725, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.168226] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.460s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.169905] env[63021]: DEBUG oslo_concurrency.lockutils [None req-13f4088b-5381-4cac-b728-cd71cf39153c tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.592s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.170112] env[63021]: DEBUG oslo_concurrency.lockutils [None req-13f4088b-5381-4cac-b728-cd71cf39153c tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.172353] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.715s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.174244] env[63021]: INFO nova.compute.claims [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 843.202131] env[63021]: INFO nova.scheduler.client.report [None req-13f4088b-5381-4cac-b728-cd71cf39153c tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Deleted allocations for instance 929cf4cd-6f3c-4988-8f80-d889d3070e1e [ 843.589229] env[63021]: DEBUG oslo_vmware.api [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293724, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064162} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.589606] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 843.590303] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-602c01d0-79d2-4c86-806c-4dd8e54abe15 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.609898] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] 6f4e6245-c0d9-46b4-baf5-5aca19e66da5/6f4e6245-c0d9-46b4-baf5-5aca19e66da5.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 843.610101] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1f7b4863-f476-4e6b-bcda-6df7fc2ed552 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.624460] env[63021]: DEBUG oslo_concurrency.lockutils [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Acquiring lock "refresh_cache-7787f6bb-e855-476f-9146-2a648932429d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.624460] env[63021]: DEBUG oslo_concurrency.lockutils [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Acquired lock "refresh_cache-7787f6bb-e855-476f-9146-2a648932429d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.624460] env[63021]: DEBUG nova.network.neutron [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 843.631722] env[63021]: DEBUG oslo_vmware.api [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Waiting for the task: (returnval){ [ 843.631722] env[63021]: value = "task-1293726" [ 843.631722] env[63021]: _type = "Task" [ 843.631722] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.645126] env[63021]: DEBUG oslo_vmware.api [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293726, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.648113] env[63021]: DEBUG oslo_vmware.api [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Task: {'id': task-1293725, 'name': PowerOnVM_Task, 'duration_secs': 0.457405} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.648360] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 843.648560] env[63021]: INFO nova.compute.manager [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Took 6.83 seconds to spawn the instance on the hypervisor. [ 843.648737] env[63021]: DEBUG nova.compute.manager [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 843.649562] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e87733c4-3afc-47fc-b2a5-80ce28bcfd52 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.670611] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Acquiring lock "4aeb790a-743b-4b58-8c38-2526cde43060" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.670876] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Lock "4aeb790a-743b-4b58-8c38-2526cde43060" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.710600] env[63021]: DEBUG oslo_concurrency.lockutils [None req-13f4088b-5381-4cac-b728-cd71cf39153c tempest-InstanceActionsNegativeTestJSON-354022710 tempest-InstanceActionsNegativeTestJSON-354022710-project-member] Lock "929cf4cd-6f3c-4988-8f80-d889d3070e1e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.540s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.142301] env[63021]: DEBUG oslo_vmware.api [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293726, 'name': ReconfigVM_Task, 'duration_secs': 0.416032} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.142577] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Reconfigured VM instance instance-0000003f to attach disk [datastore1] 6f4e6245-c0d9-46b4-baf5-5aca19e66da5/6f4e6245-c0d9-46b4-baf5-5aca19e66da5.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 844.143370] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f4049964-7155-408c-b733-98e13c8feb2b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.151807] env[63021]: DEBUG oslo_vmware.api [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Waiting for the task: (returnval){ [ 844.151807] env[63021]: value = "task-1293727" [ 844.151807] env[63021]: _type = "Task" [ 844.151807] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.164583] env[63021]: DEBUG oslo_vmware.api [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293727, 'name': Rename_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.165484] env[63021]: DEBUG nova.network.neutron [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 844.170433] env[63021]: INFO nova.compute.manager [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Took 34.08 seconds to build instance. [ 844.173907] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Lock "4aeb790a-743b-4b58-8c38-2526cde43060" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.502s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.177022] env[63021]: DEBUG nova.compute.manager [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 844.320790] env[63021]: DEBUG nova.network.neutron [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Updating instance_info_cache with network_info: [{"id": "e581fd41-04b7-4081-a7be-c6389f4d623f", "address": "fa:16:3e:5b:84:48", "network": {"id": "6d99496e-4e9f-4e5d-81a4-305f91e927cc", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-851847769-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "70e912b7ac9144c793323eaf26287ba1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f969bd9-e040-4b9b-85b2-7c61231584ad", "external-id": "nsx-vlan-transportzone-995", "segmentation_id": 995, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape581fd41-04", "ovs_interfaceid": "e581fd41-04b7-4081-a7be-c6389f4d623f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.537929] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b6c4439-bc9a-448f-8a20-34bf390de8f5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.546810] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-154730c8-becf-4591-9586-d710954c545e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.577099] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b9e4f61-bbf3-44c8-8370-7c354b75063b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.584738] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e97c7b8-9a67-4765-9cec-274bdf7a4771 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.598884] env[63021]: DEBUG nova.compute.provider_tree [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 844.663269] env[63021]: DEBUG oslo_vmware.api [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293727, 'name': Rename_Task, 'duration_secs': 0.166754} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.663552] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 844.663799] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-29f2a6b7-a7d0-4947-9b20-9f695bc38e6a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.671056] env[63021]: DEBUG oslo_vmware.api [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Waiting for the task: (returnval){ [ 844.671056] env[63021]: value = "task-1293728" [ 844.671056] env[63021]: _type = "Task" [ 844.671056] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.674851] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dbcb2059-268a-4f34-bf60-c9763a929c77 tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Lock "2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.861s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.683480] env[63021]: DEBUG nova.compute.utils [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 844.684325] env[63021]: DEBUG oslo_vmware.api [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293728, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.684943] env[63021]: DEBUG nova.compute.manager [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 844.685183] env[63021]: DEBUG nova.network.neutron [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 844.731263] env[63021]: DEBUG nova.policy [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '336467f664b647bb87a8089478e67572', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '99f669bd25374cb6969e7dca12b42476', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 844.825702] env[63021]: DEBUG oslo_concurrency.lockutils [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Releasing lock "refresh_cache-7787f6bb-e855-476f-9146-2a648932429d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.825702] env[63021]: DEBUG nova.compute.manager [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Instance network_info: |[{"id": "e581fd41-04b7-4081-a7be-c6389f4d623f", "address": "fa:16:3e:5b:84:48", "network": {"id": "6d99496e-4e9f-4e5d-81a4-305f91e927cc", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-851847769-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "70e912b7ac9144c793323eaf26287ba1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f969bd9-e040-4b9b-85b2-7c61231584ad", "external-id": "nsx-vlan-transportzone-995", "segmentation_id": 995, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape581fd41-04", "ovs_interfaceid": "e581fd41-04b7-4081-a7be-c6389f4d623f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 844.825885] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5b:84:48', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6f969bd9-e040-4b9b-85b2-7c61231584ad', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e581fd41-04b7-4081-a7be-c6389f4d623f', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 844.832191] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Creating folder: Project (70e912b7ac9144c793323eaf26287ba1). Parent ref: group-v277447. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 844.832846] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a9391371-f541-44ef-afd9-9b2a7c448436 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.845043] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Created folder: Project (70e912b7ac9144c793323eaf26287ba1) in parent group-v277447. [ 844.845271] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Creating folder: Instances. Parent ref: group-v277500. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 844.845546] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aabbfee1-840d-4468-9eea-70ea65863519 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.857396] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Created folder: Instances in parent group-v277500. [ 844.857697] env[63021]: DEBUG oslo.service.loopingcall [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 844.857985] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 844.858315] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4ea2ff43-0ef1-45cb-928a-77b4e2bc6713 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.885690] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 844.885690] env[63021]: value = "task-1293731" [ 844.885690] env[63021]: _type = "Task" [ 844.885690] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.897060] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293731, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.011241] env[63021]: DEBUG nova.network.neutron [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Successfully created port: 6d670553-cc52-4b24-8c5c-561d426a294a {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 845.105185] env[63021]: DEBUG nova.scheduler.client.report [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 845.116906] env[63021]: DEBUG nova.compute.manager [req-3c82ed90-ab20-4903-97a7-a7467254c433 req-631ec0d8-6e57-4a33-93a0-c92825ff4147 service nova] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Received event network-changed-e581fd41-04b7-4081-a7be-c6389f4d623f {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 845.117123] env[63021]: DEBUG nova.compute.manager [req-3c82ed90-ab20-4903-97a7-a7467254c433 req-631ec0d8-6e57-4a33-93a0-c92825ff4147 service nova] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Refreshing instance network info cache due to event network-changed-e581fd41-04b7-4081-a7be-c6389f4d623f. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 845.117341] env[63021]: DEBUG oslo_concurrency.lockutils [req-3c82ed90-ab20-4903-97a7-a7467254c433 req-631ec0d8-6e57-4a33-93a0-c92825ff4147 service nova] Acquiring lock "refresh_cache-7787f6bb-e855-476f-9146-2a648932429d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.117479] env[63021]: DEBUG oslo_concurrency.lockutils [req-3c82ed90-ab20-4903-97a7-a7467254c433 req-631ec0d8-6e57-4a33-93a0-c92825ff4147 service nova] Acquired lock "refresh_cache-7787f6bb-e855-476f-9146-2a648932429d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.117638] env[63021]: DEBUG nova.network.neutron [req-3c82ed90-ab20-4903-97a7-a7467254c433 req-631ec0d8-6e57-4a33-93a0-c92825ff4147 service nova] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Refreshing network info cache for port e581fd41-04b7-4081-a7be-c6389f4d623f {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 845.179329] env[63021]: DEBUG nova.compute.manager [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 845.186018] env[63021]: DEBUG oslo_vmware.api [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293728, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.189570] env[63021]: DEBUG nova.compute.manager [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 845.397440] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293731, 'name': CreateVM_Task, 'duration_secs': 0.363141} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.397567] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 845.398272] env[63021]: DEBUG oslo_concurrency.lockutils [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.398436] env[63021]: DEBUG oslo_concurrency.lockutils [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.398753] env[63021]: DEBUG oslo_concurrency.lockutils [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 845.399010] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73dbd6bf-7473-4635-93b2-a33af5ca9742 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.404270] env[63021]: DEBUG oslo_vmware.api [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Waiting for the task: (returnval){ [ 845.404270] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]521b07ba-c661-1dbd-ca62-c6b1e5b2fdbd" [ 845.404270] env[63021]: _type = "Task" [ 845.404270] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.417026] env[63021]: DEBUG oslo_vmware.api [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]521b07ba-c661-1dbd-ca62-c6b1e5b2fdbd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.620627] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.448s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.621176] env[63021]: DEBUG nova.compute.manager [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 845.623839] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.119s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.625345] env[63021]: INFO nova.compute.claims [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 845.686536] env[63021]: DEBUG oslo_vmware.api [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293728, 'name': PowerOnVM_Task, 'duration_secs': 0.5334} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.689517] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 845.689717] env[63021]: INFO nova.compute.manager [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Took 6.39 seconds to spawn the instance on the hypervisor. [ 845.689888] env[63021]: DEBUG nova.compute.manager [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 845.693877] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20d1fd5a-6766-44e7-9585-ebe32e7a7bd0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.717554] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.895481] env[63021]: DEBUG nova.network.neutron [req-3c82ed90-ab20-4903-97a7-a7467254c433 req-631ec0d8-6e57-4a33-93a0-c92825ff4147 service nova] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Updated VIF entry in instance network info cache for port e581fd41-04b7-4081-a7be-c6389f4d623f. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 845.896054] env[63021]: DEBUG nova.network.neutron [req-3c82ed90-ab20-4903-97a7-a7467254c433 req-631ec0d8-6e57-4a33-93a0-c92825ff4147 service nova] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Updating instance_info_cache with network_info: [{"id": "e581fd41-04b7-4081-a7be-c6389f4d623f", "address": "fa:16:3e:5b:84:48", "network": {"id": "6d99496e-4e9f-4e5d-81a4-305f91e927cc", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-851847769-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "70e912b7ac9144c793323eaf26287ba1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f969bd9-e040-4b9b-85b2-7c61231584ad", "external-id": "nsx-vlan-transportzone-995", "segmentation_id": 995, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape581fd41-04", "ovs_interfaceid": "e581fd41-04b7-4081-a7be-c6389f4d623f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.916480] env[63021]: DEBUG oslo_vmware.api [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]521b07ba-c661-1dbd-ca62-c6b1e5b2fdbd, 'name': SearchDatastore_Task, 'duration_secs': 0.011787} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.916918] env[63021]: DEBUG oslo_concurrency.lockutils [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.917081] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 845.917343] env[63021]: DEBUG oslo_concurrency.lockutils [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.917481] env[63021]: DEBUG oslo_concurrency.lockutils [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.917664] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 845.918590] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2db58782-955f-42a2-96b3-76376e7adfc7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.930720] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 845.930940] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 845.931791] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c78c060-52b7-4a56-a557-1ea19b38c36a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.942747] env[63021]: DEBUG oslo_vmware.api [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Waiting for the task: (returnval){ [ 845.942747] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5221c1f7-e18b-e2f8-bdb5-5370d219f9c1" [ 845.942747] env[63021]: _type = "Task" [ 845.942747] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.958582] env[63021]: DEBUG oslo_vmware.api [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5221c1f7-e18b-e2f8-bdb5-5370d219f9c1, 'name': SearchDatastore_Task, 'duration_secs': 0.011367} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.960481] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a3f48f1-8533-4679-978c-6d8ae4a45d0a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.970746] env[63021]: DEBUG oslo_vmware.api [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Waiting for the task: (returnval){ [ 845.970746] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]523831de-338b-0fc4-f71a-e31a2cb1f4c2" [ 845.970746] env[63021]: _type = "Task" [ 845.970746] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.980980] env[63021]: DEBUG oslo_vmware.api [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]523831de-338b-0fc4-f71a-e31a2cb1f4c2, 'name': SearchDatastore_Task, 'duration_secs': 0.010032} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.982242] env[63021]: DEBUG oslo_concurrency.lockutils [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.982322] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 7787f6bb-e855-476f-9146-2a648932429d/7787f6bb-e855-476f-9146-2a648932429d.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 845.982532] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-876edbd1-65e5-488d-bb89-c4a59d8717af {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.991656] env[63021]: DEBUG oslo_vmware.api [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Waiting for the task: (returnval){ [ 845.991656] env[63021]: value = "task-1293732" [ 845.991656] env[63021]: _type = "Task" [ 845.991656] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.000955] env[63021]: DEBUG oslo_vmware.api [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': task-1293732, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.130774] env[63021]: DEBUG nova.compute.utils [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 846.132388] env[63021]: DEBUG nova.compute.manager [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 846.132569] env[63021]: DEBUG nova.network.neutron [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 846.192086] env[63021]: DEBUG nova.policy [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2642402689b04f89946802ac3b3e6287', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '805af233efa147b3982fb794cd9e6f73', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 846.201982] env[63021]: DEBUG nova.compute.manager [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 846.226487] env[63021]: INFO nova.compute.manager [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Took 29.89 seconds to build instance. [ 846.246028] env[63021]: DEBUG nova.virt.hardware [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 846.246406] env[63021]: DEBUG nova.virt.hardware [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 846.246507] env[63021]: DEBUG nova.virt.hardware [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 846.246690] env[63021]: DEBUG nova.virt.hardware [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 846.246838] env[63021]: DEBUG nova.virt.hardware [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 846.246982] env[63021]: DEBUG nova.virt.hardware [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 846.247212] env[63021]: DEBUG nova.virt.hardware [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 846.247375] env[63021]: DEBUG nova.virt.hardware [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 846.247591] env[63021]: DEBUG nova.virt.hardware [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 846.247775] env[63021]: DEBUG nova.virt.hardware [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 846.247959] env[63021]: DEBUG nova.virt.hardware [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 846.248882] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61cf5ded-d614-4105-a471-a3dc942b7d96 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.262097] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72154404-0f30-4480-ac6c-773ab369659c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.398504] env[63021]: DEBUG oslo_concurrency.lockutils [req-3c82ed90-ab20-4903-97a7-a7467254c433 req-631ec0d8-6e57-4a33-93a0-c92825ff4147 service nova] Releasing lock "refresh_cache-7787f6bb-e855-476f-9146-2a648932429d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.503635] env[63021]: DEBUG oslo_vmware.api [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': task-1293732, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.484355} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.503635] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 7787f6bb-e855-476f-9146-2a648932429d/7787f6bb-e855-476f-9146-2a648932429d.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 846.503635] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 846.503815] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5eed54db-1dfe-4b5f-8ab9-4fbe2b550544 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.511985] env[63021]: DEBUG oslo_vmware.api [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Waiting for the task: (returnval){ [ 846.511985] env[63021]: value = "task-1293733" [ 846.511985] env[63021]: _type = "Task" [ 846.511985] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.522914] env[63021]: DEBUG oslo_vmware.api [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': task-1293733, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.529112] env[63021]: DEBUG oslo_concurrency.lockutils [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Acquiring lock "80b857c4-d9dd-4483-970f-a0f5f4bcf173" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.529555] env[63021]: DEBUG oslo_concurrency.lockutils [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Lock "80b857c4-d9dd-4483-970f-a0f5f4bcf173" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.642025] env[63021]: DEBUG nova.compute.manager [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 846.651365] env[63021]: DEBUG nova.network.neutron [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Successfully created port: 1019be13-e855-4eab-8201-a2ae2b22476c {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 846.728702] env[63021]: DEBUG oslo_concurrency.lockutils [None req-990a87cb-26e6-4caa-b18d-976aef8ec358 tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Lock "6f4e6245-c0d9-46b4-baf5-5aca19e66da5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.724s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.817070] env[63021]: DEBUG nova.network.neutron [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Successfully updated port: 6d670553-cc52-4b24-8c5c-561d426a294a {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 846.836162] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Acquiring lock "2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.836338] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Lock "2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.836601] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Acquiring lock "2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.836791] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Lock "2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.836956] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Lock "2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.841788] env[63021]: INFO nova.compute.manager [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Terminating instance [ 846.846765] env[63021]: DEBUG nova.compute.manager [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 846.846953] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 846.848819] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5926e354-f8df-492b-bf8e-823cda5feff9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.857862] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 846.860679] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b3e86296-533a-4d55-957f-c67189c0db82 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.862400] env[63021]: INFO nova.compute.manager [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Rebuilding instance [ 846.872571] env[63021]: DEBUG oslo_vmware.api [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Waiting for the task: (returnval){ [ 846.872571] env[63021]: value = "task-1293734" [ 846.872571] env[63021]: _type = "Task" [ 846.872571] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.883062] env[63021]: DEBUG oslo_vmware.api [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Task: {'id': task-1293734, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.930961] env[63021]: DEBUG nova.compute.manager [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 846.931502] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfcf0cad-d0c0-454a-83c3-fabd2138310d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.026606] env[63021]: DEBUG oslo_vmware.api [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': task-1293733, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078445} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.026924] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 847.027901] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d5b8d62-0cf7-4d2a-97b4-5fc66797d0f9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.035020] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a36ba827-036b-49a4-91a8-1466860b237b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.059269] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] 7787f6bb-e855-476f-9146-2a648932429d/7787f6bb-e855-476f-9146-2a648932429d.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 847.060393] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-23145af8-39ad-4111-b87d-6d3d61537e22 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.079018] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd1702bc-e465-4696-8173-721e89a5f08a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.085467] env[63021]: DEBUG oslo_vmware.api [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Waiting for the task: (returnval){ [ 847.085467] env[63021]: value = "task-1293735" [ 847.085467] env[63021]: _type = "Task" [ 847.085467] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.124566] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c8945a4-432b-4642-a6fb-b766e8e766f5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.127359] env[63021]: DEBUG oslo_vmware.api [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': task-1293735, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.134276] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a776cf2f-0c09-44e4-b20a-aa670642b891 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.156027] env[63021]: DEBUG nova.compute.provider_tree [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 847.232049] env[63021]: DEBUG nova.compute.manager [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 847.242701] env[63021]: DEBUG nova.compute.manager [req-bf93df79-c753-4e9b-9fbf-1ef1f0f8c0dd req-1f5e1712-e17d-4e2c-b15c-421fe75b7da8 service nova] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Received event network-vif-plugged-6d670553-cc52-4b24-8c5c-561d426a294a {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 847.242916] env[63021]: DEBUG oslo_concurrency.lockutils [req-bf93df79-c753-4e9b-9fbf-1ef1f0f8c0dd req-1f5e1712-e17d-4e2c-b15c-421fe75b7da8 service nova] Acquiring lock "70f67905-7c60-433d-9ebc-d66fa44eb36e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.243139] env[63021]: DEBUG oslo_concurrency.lockutils [req-bf93df79-c753-4e9b-9fbf-1ef1f0f8c0dd req-1f5e1712-e17d-4e2c-b15c-421fe75b7da8 service nova] Lock "70f67905-7c60-433d-9ebc-d66fa44eb36e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.243311] env[63021]: DEBUG oslo_concurrency.lockutils [req-bf93df79-c753-4e9b-9fbf-1ef1f0f8c0dd req-1f5e1712-e17d-4e2c-b15c-421fe75b7da8 service nova] Lock "70f67905-7c60-433d-9ebc-d66fa44eb36e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.243515] env[63021]: DEBUG nova.compute.manager [req-bf93df79-c753-4e9b-9fbf-1ef1f0f8c0dd req-1f5e1712-e17d-4e2c-b15c-421fe75b7da8 service nova] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] No waiting events found dispatching network-vif-plugged-6d670553-cc52-4b24-8c5c-561d426a294a {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 847.243700] env[63021]: WARNING nova.compute.manager [req-bf93df79-c753-4e9b-9fbf-1ef1f0f8c0dd req-1f5e1712-e17d-4e2c-b15c-421fe75b7da8 service nova] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Received unexpected event network-vif-plugged-6d670553-cc52-4b24-8c5c-561d426a294a for instance with vm_state building and task_state spawning. [ 847.243865] env[63021]: DEBUG nova.compute.manager [req-bf93df79-c753-4e9b-9fbf-1ef1f0f8c0dd req-1f5e1712-e17d-4e2c-b15c-421fe75b7da8 service nova] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Received event network-changed-6d670553-cc52-4b24-8c5c-561d426a294a {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 847.244028] env[63021]: DEBUG nova.compute.manager [req-bf93df79-c753-4e9b-9fbf-1ef1f0f8c0dd req-1f5e1712-e17d-4e2c-b15c-421fe75b7da8 service nova] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Refreshing instance network info cache due to event network-changed-6d670553-cc52-4b24-8c5c-561d426a294a. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 847.244297] env[63021]: DEBUG oslo_concurrency.lockutils [req-bf93df79-c753-4e9b-9fbf-1ef1f0f8c0dd req-1f5e1712-e17d-4e2c-b15c-421fe75b7da8 service nova] Acquiring lock "refresh_cache-70f67905-7c60-433d-9ebc-d66fa44eb36e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.244450] env[63021]: DEBUG oslo_concurrency.lockutils [req-bf93df79-c753-4e9b-9fbf-1ef1f0f8c0dd req-1f5e1712-e17d-4e2c-b15c-421fe75b7da8 service nova] Acquired lock "refresh_cache-70f67905-7c60-433d-9ebc-d66fa44eb36e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.244581] env[63021]: DEBUG nova.network.neutron [req-bf93df79-c753-4e9b-9fbf-1ef1f0f8c0dd req-1f5e1712-e17d-4e2c-b15c-421fe75b7da8 service nova] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Refreshing network info cache for port 6d670553-cc52-4b24-8c5c-561d426a294a {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 847.320665] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Acquiring lock "refresh_cache-70f67905-7c60-433d-9ebc-d66fa44eb36e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.382305] env[63021]: DEBUG oslo_vmware.api [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Task: {'id': task-1293734, 'name': PowerOffVM_Task, 'duration_secs': 0.247192} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.382578] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 847.383327] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 847.383327] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-13dc7bac-3df3-4466-9f7f-9237edfd40ca {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.445708] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 847.446670] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5e3d98af-73a6-491f-b2b9-2d4d66ed0e60 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.449238] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 847.450041] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 847.450041] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Deleting the datastore file [datastore1] 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4 {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 847.450198] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fbc6b275-cb9a-4694-bf6d-fbf61dbaea72 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.456243] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Waiting for the task: (returnval){ [ 847.456243] env[63021]: value = "task-1293737" [ 847.456243] env[63021]: _type = "Task" [ 847.456243] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.457627] env[63021]: DEBUG oslo_vmware.api [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Waiting for the task: (returnval){ [ 847.457627] env[63021]: value = "task-1293738" [ 847.457627] env[63021]: _type = "Task" [ 847.457627] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.469293] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293737, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.472454] env[63021]: DEBUG oslo_vmware.api [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Task: {'id': task-1293738, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.596537] env[63021]: DEBUG oslo_vmware.api [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': task-1293735, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.656246] env[63021]: DEBUG nova.compute.manager [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 847.659073] env[63021]: DEBUG nova.scheduler.client.report [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 847.683054] env[63021]: DEBUG nova.virt.hardware [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 847.683054] env[63021]: DEBUG nova.virt.hardware [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 847.683246] env[63021]: DEBUG nova.virt.hardware [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 847.683438] env[63021]: DEBUG nova.virt.hardware [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 847.683586] env[63021]: DEBUG nova.virt.hardware [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 847.683739] env[63021]: DEBUG nova.virt.hardware [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 847.683955] env[63021]: DEBUG nova.virt.hardware [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 847.684190] env[63021]: DEBUG nova.virt.hardware [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 847.684379] env[63021]: DEBUG nova.virt.hardware [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 847.684544] env[63021]: DEBUG nova.virt.hardware [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 847.684713] env[63021]: DEBUG nova.virt.hardware [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 847.685846] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b460902-d943-4c46-bafd-aa857d8e3cde {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.695641] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff752b16-2395-4e11-aff1-ed5a8320d220 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.751868] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.781187] env[63021]: DEBUG nova.network.neutron [req-bf93df79-c753-4e9b-9fbf-1ef1f0f8c0dd req-1f5e1712-e17d-4e2c-b15c-421fe75b7da8 service nova] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 847.859609] env[63021]: DEBUG nova.network.neutron [req-bf93df79-c753-4e9b-9fbf-1ef1f0f8c0dd req-1f5e1712-e17d-4e2c-b15c-421fe75b7da8 service nova] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.974344] env[63021]: DEBUG oslo_vmware.api [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Task: {'id': task-1293738, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14017} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.978106] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 847.978310] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 847.978492] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 847.978661] env[63021]: INFO nova.compute.manager [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Took 1.13 seconds to destroy the instance on the hypervisor. [ 847.978906] env[63021]: DEBUG oslo.service.loopingcall [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 847.979130] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293737, 'name': PowerOffVM_Task, 'duration_secs': 0.159002} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.979420] env[63021]: DEBUG nova.compute.manager [-] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 847.979578] env[63021]: DEBUG nova.network.neutron [-] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 847.982197] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 847.983198] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 847.983992] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1c5057e-8c4e-4745-b76f-3eb37ffb0121 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.991303] env[63021]: DEBUG oslo_concurrency.lockutils [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Acquiring lock "4d712cc1-6202-407e-8767-c620a1ce0bba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.991517] env[63021]: DEBUG oslo_concurrency.lockutils [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Lock "4d712cc1-6202-407e-8767-c620a1ce0bba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.996640] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 847.996853] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8c66d1c5-894a-4b14-923f-22ba5a40d7f9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.029783] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 848.030100] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 848.030194] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Deleting the datastore file [datastore1] 6f4e6245-c0d9-46b4-baf5-5aca19e66da5 {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 848.030454] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a9b9b52a-82df-47cf-9591-58e915fb23ab {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.037745] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Waiting for the task: (returnval){ [ 848.037745] env[63021]: value = "task-1293740" [ 848.037745] env[63021]: _type = "Task" [ 848.037745] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.046150] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293740, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.097748] env[63021]: DEBUG oslo_vmware.api [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': task-1293735, 'name': ReconfigVM_Task, 'duration_secs': 0.654652} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.098196] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Reconfigured VM instance instance-00000041 to attach disk [datastore1] 7787f6bb-e855-476f-9146-2a648932429d/7787f6bb-e855-476f-9146-2a648932429d.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 848.098763] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-708d49a0-189a-4bbf-9a81-4cf34dffd30d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.106675] env[63021]: DEBUG oslo_vmware.api [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Waiting for the task: (returnval){ [ 848.106675] env[63021]: value = "task-1293741" [ 848.106675] env[63021]: _type = "Task" [ 848.106675] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.115681] env[63021]: DEBUG oslo_vmware.api [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': task-1293741, 'name': Rename_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.164777] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.541s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.165350] env[63021]: DEBUG nova.compute.manager [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 848.168238] env[63021]: DEBUG oslo_concurrency.lockutils [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.880s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.168482] env[63021]: DEBUG nova.objects.instance [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Lazy-loading 'resources' on Instance uuid bef5c7cb-be2f-4be7-b72c-4a066d473dd0 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 848.361947] env[63021]: DEBUG oslo_concurrency.lockutils [req-bf93df79-c753-4e9b-9fbf-1ef1f0f8c0dd req-1f5e1712-e17d-4e2c-b15c-421fe75b7da8 service nova] Releasing lock "refresh_cache-70f67905-7c60-433d-9ebc-d66fa44eb36e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.362363] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Acquired lock "refresh_cache-70f67905-7c60-433d-9ebc-d66fa44eb36e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.362524] env[63021]: DEBUG nova.network.neutron [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 848.427855] env[63021]: DEBUG nova.network.neutron [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Successfully updated port: 1019be13-e855-4eab-8201-a2ae2b22476c {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 848.547597] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293740, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.12229} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.547851] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 848.548040] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 848.548220] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 848.618933] env[63021]: DEBUG oslo_vmware.api [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': task-1293741, 'name': Rename_Task, 'duration_secs': 0.172057} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.619333] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 848.619643] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5c81483f-3284-4dc8-a42c-30270efba1bb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.627413] env[63021]: DEBUG oslo_vmware.api [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Waiting for the task: (returnval){ [ 848.627413] env[63021]: value = "task-1293742" [ 848.627413] env[63021]: _type = "Task" [ 848.627413] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.635314] env[63021]: DEBUG oslo_vmware.api [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': task-1293742, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.674364] env[63021]: DEBUG nova.compute.utils [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 848.677044] env[63021]: DEBUG nova.compute.manager [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 848.677044] env[63021]: DEBUG nova.network.neutron [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 848.723051] env[63021]: DEBUG nova.policy [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c8c1392321694d17bc08633dfbfe5ee2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5661c44b534b4a7e9452e914fc80d66e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 848.860762] env[63021]: DEBUG nova.network.neutron [-] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.914155] env[63021]: DEBUG nova.network.neutron [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 848.930821] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "refresh_cache-71909213-258d-45f5-9a3e-e473e8ce8aa2" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.931083] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquired lock "refresh_cache-71909213-258d-45f5-9a3e-e473e8ce8aa2" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.931208] env[63021]: DEBUG nova.network.neutron [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 849.015975] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-199641df-6bc8-410a-8d29-96b8cefa23d2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.026388] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d8df768-bb65-4d5b-9ad1-cb5052749676 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.063300] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fdf33d2-efc9-42c5-8012-bd5aaf0f3f80 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.072971] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddc895a0-928a-452a-8b85-8c48f5e4b4d9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.089093] env[63021]: DEBUG nova.compute.provider_tree [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 849.139313] env[63021]: DEBUG oslo_vmware.api [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': task-1293742, 'name': PowerOnVM_Task, 'duration_secs': 0.449965} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.139668] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 849.139874] env[63021]: INFO nova.compute.manager [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Took 6.41 seconds to spawn the instance on the hypervisor. [ 849.140067] env[63021]: DEBUG nova.compute.manager [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 849.140982] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d12c997f-0ced-47e7-ac64-7561a2114125 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.147310] env[63021]: DEBUG nova.network.neutron [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Successfully created port: a43dcfd2-397e-44fe-bd1c-b6116c25a28a {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 849.177420] env[63021]: DEBUG nova.compute.manager [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 849.210441] env[63021]: DEBUG nova.network.neutron [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Updating instance_info_cache with network_info: [{"id": "6d670553-cc52-4b24-8c5c-561d426a294a", "address": "fa:16:3e:fb:65:2d", "network": {"id": "d6eb7ea2-5a1c-46b5-a443-c265c4eeba2e", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-498672604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "99f669bd25374cb6969e7dca12b42476", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f39e3b37-7906-4bbc-820e-ceac74e4d827", "external-id": "nsx-vlan-transportzone-328", "segmentation_id": 328, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d670553-cc", "ovs_interfaceid": "6d670553-cc52-4b24-8c5c-561d426a294a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.320770] env[63021]: DEBUG nova.compute.manager [req-8b09fffc-4545-483c-8843-41be8815dbb5 req-fa9104c0-6671-4488-868b-5fe7b09308d1 service nova] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Received event network-vif-plugged-1019be13-e855-4eab-8201-a2ae2b22476c {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 849.320939] env[63021]: DEBUG oslo_concurrency.lockutils [req-8b09fffc-4545-483c-8843-41be8815dbb5 req-fa9104c0-6671-4488-868b-5fe7b09308d1 service nova] Acquiring lock "71909213-258d-45f5-9a3e-e473e8ce8aa2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.321187] env[63021]: DEBUG oslo_concurrency.lockutils [req-8b09fffc-4545-483c-8843-41be8815dbb5 req-fa9104c0-6671-4488-868b-5fe7b09308d1 service nova] Lock "71909213-258d-45f5-9a3e-e473e8ce8aa2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.321362] env[63021]: DEBUG oslo_concurrency.lockutils [req-8b09fffc-4545-483c-8843-41be8815dbb5 req-fa9104c0-6671-4488-868b-5fe7b09308d1 service nova] Lock "71909213-258d-45f5-9a3e-e473e8ce8aa2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.321515] env[63021]: DEBUG nova.compute.manager [req-8b09fffc-4545-483c-8843-41be8815dbb5 req-fa9104c0-6671-4488-868b-5fe7b09308d1 service nova] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] No waiting events found dispatching network-vif-plugged-1019be13-e855-4eab-8201-a2ae2b22476c {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 849.321670] env[63021]: WARNING nova.compute.manager [req-8b09fffc-4545-483c-8843-41be8815dbb5 req-fa9104c0-6671-4488-868b-5fe7b09308d1 service nova] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Received unexpected event network-vif-plugged-1019be13-e855-4eab-8201-a2ae2b22476c for instance with vm_state building and task_state spawning. [ 849.321820] env[63021]: DEBUG nova.compute.manager [req-8b09fffc-4545-483c-8843-41be8815dbb5 req-fa9104c0-6671-4488-868b-5fe7b09308d1 service nova] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Received event network-vif-deleted-d82f4ff7-a0f2-401a-8b11-1fdcc8ba0bc8 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 849.322105] env[63021]: DEBUG nova.compute.manager [req-8b09fffc-4545-483c-8843-41be8815dbb5 req-fa9104c0-6671-4488-868b-5fe7b09308d1 service nova] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Received event network-changed-1019be13-e855-4eab-8201-a2ae2b22476c {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 849.322305] env[63021]: DEBUG nova.compute.manager [req-8b09fffc-4545-483c-8843-41be8815dbb5 req-fa9104c0-6671-4488-868b-5fe7b09308d1 service nova] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Refreshing instance network info cache due to event network-changed-1019be13-e855-4eab-8201-a2ae2b22476c. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 849.322474] env[63021]: DEBUG oslo_concurrency.lockutils [req-8b09fffc-4545-483c-8843-41be8815dbb5 req-fa9104c0-6671-4488-868b-5fe7b09308d1 service nova] Acquiring lock "refresh_cache-71909213-258d-45f5-9a3e-e473e8ce8aa2" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.363516] env[63021]: INFO nova.compute.manager [-] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Took 1.38 seconds to deallocate network for instance. [ 849.460523] env[63021]: DEBUG nova.network.neutron [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 849.580696] env[63021]: DEBUG nova.network.neutron [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Updating instance_info_cache with network_info: [{"id": "1019be13-e855-4eab-8201-a2ae2b22476c", "address": "fa:16:3e:32:70:89", "network": {"id": "708d0dc3-c651-4001-9638-1d3538ff76c9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-983471056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "805af233efa147b3982fb794cd9e6f73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1019be13-e8", "ovs_interfaceid": "1019be13-e855-4eab-8201-a2ae2b22476c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.593250] env[63021]: DEBUG nova.virt.hardware [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 849.593689] env[63021]: DEBUG nova.virt.hardware [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 849.593954] env[63021]: DEBUG nova.virt.hardware [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 849.594601] env[63021]: DEBUG nova.virt.hardware [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 849.594601] env[63021]: DEBUG nova.virt.hardware [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 849.594601] env[63021]: DEBUG nova.virt.hardware [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 849.594792] env[63021]: DEBUG nova.virt.hardware [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 849.594953] env[63021]: DEBUG nova.virt.hardware [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 849.595142] env[63021]: DEBUG nova.virt.hardware [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 849.595309] env[63021]: DEBUG nova.virt.hardware [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 849.595478] env[63021]: DEBUG nova.virt.hardware [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 849.596318] env[63021]: DEBUG nova.scheduler.client.report [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 849.599773] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd7c25da-6071-4e91-a7b4-7d272c5d6aa2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.609350] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3db4ba9e-b129-4b8f-b707-e454fb3cd239 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.624297] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Instance VIF info [] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 849.629700] env[63021]: DEBUG oslo.service.loopingcall [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 849.630186] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 849.630400] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e3f60080-9512-44c2-8a0e-839edbcef2c9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.648375] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 849.648375] env[63021]: value = "task-1293743" [ 849.648375] env[63021]: _type = "Task" [ 849.648375] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.660613] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293743, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.669180] env[63021]: INFO nova.compute.manager [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Took 30.53 seconds to build instance. [ 849.712686] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Releasing lock "refresh_cache-70f67905-7c60-433d-9ebc-d66fa44eb36e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.713053] env[63021]: DEBUG nova.compute.manager [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Instance network_info: |[{"id": "6d670553-cc52-4b24-8c5c-561d426a294a", "address": "fa:16:3e:fb:65:2d", "network": {"id": "d6eb7ea2-5a1c-46b5-a443-c265c4eeba2e", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-498672604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "99f669bd25374cb6969e7dca12b42476", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f39e3b37-7906-4bbc-820e-ceac74e4d827", "external-id": "nsx-vlan-transportzone-328", "segmentation_id": 328, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d670553-cc", "ovs_interfaceid": "6d670553-cc52-4b24-8c5c-561d426a294a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 849.713482] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fb:65:2d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f39e3b37-7906-4bbc-820e-ceac74e4d827', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6d670553-cc52-4b24-8c5c-561d426a294a', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 849.721080] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Creating folder: Project (99f669bd25374cb6969e7dca12b42476). Parent ref: group-v277447. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 849.722087] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-131fdfb5-9816-45c8-87ae-723543e984d8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.735892] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Created folder: Project (99f669bd25374cb6969e7dca12b42476) in parent group-v277447. [ 849.736108] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Creating folder: Instances. Parent ref: group-v277504. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 849.736364] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9feb4f0b-4b22-4ca7-8dab-fd28e3b72f83 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.748729] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Created folder: Instances in parent group-v277504. [ 849.748977] env[63021]: DEBUG oslo.service.loopingcall [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 849.749180] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 849.749393] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-595e4400-8320-4478-9b64-def35cea4379 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.769697] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 849.769697] env[63021]: value = "task-1293746" [ 849.769697] env[63021]: _type = "Task" [ 849.769697] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.781672] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293746, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.870591] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.085390] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Releasing lock "refresh_cache-71909213-258d-45f5-9a3e-e473e8ce8aa2" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.085744] env[63021]: DEBUG nova.compute.manager [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Instance network_info: |[{"id": "1019be13-e855-4eab-8201-a2ae2b22476c", "address": "fa:16:3e:32:70:89", "network": {"id": "708d0dc3-c651-4001-9638-1d3538ff76c9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-983471056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "805af233efa147b3982fb794cd9e6f73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1019be13-e8", "ovs_interfaceid": "1019be13-e855-4eab-8201-a2ae2b22476c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 850.086061] env[63021]: DEBUG oslo_concurrency.lockutils [req-8b09fffc-4545-483c-8843-41be8815dbb5 req-fa9104c0-6671-4488-868b-5fe7b09308d1 service nova] Acquired lock "refresh_cache-71909213-258d-45f5-9a3e-e473e8ce8aa2" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.086243] env[63021]: DEBUG nova.network.neutron [req-8b09fffc-4545-483c-8843-41be8815dbb5 req-fa9104c0-6671-4488-868b-5fe7b09308d1 service nova] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Refreshing network info cache for port 1019be13-e855-4eab-8201-a2ae2b22476c {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 850.087634] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:32:70:89', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b7a73c01-1bb9-4612-a1a7-16d71b732e81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1019be13-e855-4eab-8201-a2ae2b22476c', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 850.099629] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Creating folder: Project (805af233efa147b3982fb794cd9e6f73). Parent ref: group-v277447. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 850.102380] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-daa438c7-1a7a-47b7-93e5-03ced524979e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.105467] env[63021]: DEBUG oslo_concurrency.lockutils [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.937s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.108174] env[63021]: DEBUG oslo_concurrency.lockutils [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.285s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.111139] env[63021]: INFO nova.compute.claims [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 850.127651] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Created folder: Project (805af233efa147b3982fb794cd9e6f73) in parent group-v277447. [ 850.127651] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Creating folder: Instances. Parent ref: group-v277507. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 850.129598] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-02707e96-ada0-4390-a287-6f1ae96a8b05 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.139492] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Created folder: Instances in parent group-v277507. [ 850.139776] env[63021]: DEBUG oslo.service.loopingcall [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 850.139915] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 850.140139] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-29bfd025-d736-4c88-9353-91f1be12028d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.161578] env[63021]: INFO nova.scheduler.client.report [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Deleted allocations for instance bef5c7cb-be2f-4be7-b72c-4a066d473dd0 [ 850.175417] env[63021]: DEBUG oslo_concurrency.lockutils [None req-34d97618-1fa1-47e1-9630-a1562aab0ecd tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Lock "7787f6bb-e855-476f-9146-2a648932429d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 97.639s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.181793] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 850.181793] env[63021]: value = "task-1293749" [ 850.181793] env[63021]: _type = "Task" [ 850.181793] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.185361] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293743, 'name': CreateVM_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.189967] env[63021]: DEBUG nova.compute.manager [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 850.199585] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293749, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.221445] env[63021]: DEBUG nova.virt.hardware [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 850.221722] env[63021]: DEBUG nova.virt.hardware [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 850.221917] env[63021]: DEBUG nova.virt.hardware [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 850.222178] env[63021]: DEBUG nova.virt.hardware [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 850.222365] env[63021]: DEBUG nova.virt.hardware [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 850.222542] env[63021]: DEBUG nova.virt.hardware [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 850.222784] env[63021]: DEBUG nova.virt.hardware [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 850.222991] env[63021]: DEBUG nova.virt.hardware [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 850.223238] env[63021]: DEBUG nova.virt.hardware [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 850.223451] env[63021]: DEBUG nova.virt.hardware [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 850.223681] env[63021]: DEBUG nova.virt.hardware [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 850.224607] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a92dcf2d-69b8-4372-abf9-423a8154b65c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.233318] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3208a802-f702-4835-b40b-dc7075866758 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.280113] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293746, 'name': CreateVM_Task, 'duration_secs': 0.294548} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.282792] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 850.283494] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.283659] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.283992] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 850.284572] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38ba2bb4-13b9-4b4d-99f6-fbc06efdf710 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.290043] env[63021]: DEBUG oslo_vmware.api [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Waiting for the task: (returnval){ [ 850.290043] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52bf5bd1-d305-ee9a-ec3c-89b7894d6b7a" [ 850.290043] env[63021]: _type = "Task" [ 850.290043] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.299101] env[63021]: DEBUG oslo_vmware.api [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52bf5bd1-d305-ee9a-ec3c-89b7894d6b7a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.382733] env[63021]: DEBUG nova.network.neutron [req-8b09fffc-4545-483c-8843-41be8815dbb5 req-fa9104c0-6671-4488-868b-5fe7b09308d1 service nova] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Updated VIF entry in instance network info cache for port 1019be13-e855-4eab-8201-a2ae2b22476c. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 850.383104] env[63021]: DEBUG nova.network.neutron [req-8b09fffc-4545-483c-8843-41be8815dbb5 req-fa9104c0-6671-4488-868b-5fe7b09308d1 service nova] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Updating instance_info_cache with network_info: [{"id": "1019be13-e855-4eab-8201-a2ae2b22476c", "address": "fa:16:3e:32:70:89", "network": {"id": "708d0dc3-c651-4001-9638-1d3538ff76c9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-983471056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "805af233efa147b3982fb794cd9e6f73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1019be13-e8", "ovs_interfaceid": "1019be13-e855-4eab-8201-a2ae2b22476c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.659244] env[63021]: DEBUG nova.network.neutron [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Successfully updated port: a43dcfd2-397e-44fe-bd1c-b6116c25a28a {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 850.677037] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293743, 'name': CreateVM_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.678554] env[63021]: INFO nova.compute.manager [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Rescuing [ 850.678778] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Acquiring lock "refresh_cache-7787f6bb-e855-476f-9146-2a648932429d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.678922] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Acquired lock "refresh_cache-7787f6bb-e855-476f-9146-2a648932429d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.679123] env[63021]: DEBUG nova.network.neutron [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 850.682351] env[63021]: DEBUG nova.compute.manager [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 850.684870] env[63021]: DEBUG oslo_concurrency.lockutils [None req-703dca1f-12b2-4996-954e-fa1dd5828a71 tempest-ServerMetadataNegativeTestJSON-880021035 tempest-ServerMetadataNegativeTestJSON-880021035-project-member] Lock "bef5c7cb-be2f-4be7-b72c-4a066d473dd0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.883s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.696556] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293749, 'name': CreateVM_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.800590] env[63021]: DEBUG oslo_vmware.api [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52bf5bd1-d305-ee9a-ec3c-89b7894d6b7a, 'name': SearchDatastore_Task, 'duration_secs': 0.013812} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.801212] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.801592] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 850.801954] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.802244] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.802622] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 850.803047] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b0d0395d-f1db-4aed-8229-0e56d5546b7b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.812621] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 850.812933] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 850.813786] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a0cdacd-ff59-41aa-b064-c2b4bfde8500 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.819673] env[63021]: DEBUG oslo_vmware.api [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Waiting for the task: (returnval){ [ 850.819673] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]529ee228-7d33-723c-85e2-9ffb9adb63d5" [ 850.819673] env[63021]: _type = "Task" [ 850.819673] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.829084] env[63021]: DEBUG oslo_vmware.api [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]529ee228-7d33-723c-85e2-9ffb9adb63d5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.886820] env[63021]: DEBUG oslo_concurrency.lockutils [req-8b09fffc-4545-483c-8843-41be8815dbb5 req-fa9104c0-6671-4488-868b-5fe7b09308d1 service nova] Releasing lock "refresh_cache-71909213-258d-45f5-9a3e-e473e8ce8aa2" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.162183] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquiring lock "refresh_cache-df856146-61ba-4de7-a5fd-d75263927163" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.162183] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquired lock "refresh_cache-df856146-61ba-4de7-a5fd-d75263927163" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.162183] env[63021]: DEBUG nova.network.neutron [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 851.185395] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293743, 'name': CreateVM_Task, 'duration_secs': 1.321129} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.185656] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 851.195154] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.195373] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.195714] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 851.200691] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fcb69390-7903-45d0-939d-33ba86c8a473 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.210456] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293749, 'name': CreateVM_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.210753] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Waiting for the task: (returnval){ [ 851.210753] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5285bae7-d36b-4e8e-a4e8-b07231e55670" [ 851.210753] env[63021]: _type = "Task" [ 851.210753] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.218200] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.223722] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5285bae7-d36b-4e8e-a4e8-b07231e55670, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.330494] env[63021]: DEBUG oslo_vmware.api [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]529ee228-7d33-723c-85e2-9ffb9adb63d5, 'name': SearchDatastore_Task, 'duration_secs': 0.010157} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.333487] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84e2f5b6-00c9-44fb-a05d-4a4f0252b413 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.339831] env[63021]: DEBUG oslo_vmware.api [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Waiting for the task: (returnval){ [ 851.339831] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5215407b-195c-0045-0c3c-b4a147e486c6" [ 851.339831] env[63021]: _type = "Task" [ 851.339831] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.348313] env[63021]: DEBUG oslo_vmware.api [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5215407b-195c-0045-0c3c-b4a147e486c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.386822] env[63021]: DEBUG nova.network.neutron [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Updating instance_info_cache with network_info: [{"id": "e581fd41-04b7-4081-a7be-c6389f4d623f", "address": "fa:16:3e:5b:84:48", "network": {"id": "6d99496e-4e9f-4e5d-81a4-305f91e927cc", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-851847769-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "70e912b7ac9144c793323eaf26287ba1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f969bd9-e040-4b9b-85b2-7c61231584ad", "external-id": "nsx-vlan-transportzone-995", "segmentation_id": 995, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape581fd41-04", "ovs_interfaceid": "e581fd41-04b7-4081-a7be-c6389f4d623f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.487305] env[63021]: DEBUG nova.compute.manager [req-021358fd-6bee-4588-91ff-74d563232536 req-64356039-492d-4bbb-85e6-87dbe94f364e service nova] [instance: df856146-61ba-4de7-a5fd-d75263927163] Received event network-vif-plugged-a43dcfd2-397e-44fe-bd1c-b6116c25a28a {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 851.487526] env[63021]: DEBUG oslo_concurrency.lockutils [req-021358fd-6bee-4588-91ff-74d563232536 req-64356039-492d-4bbb-85e6-87dbe94f364e service nova] Acquiring lock "df856146-61ba-4de7-a5fd-d75263927163-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.487735] env[63021]: DEBUG oslo_concurrency.lockutils [req-021358fd-6bee-4588-91ff-74d563232536 req-64356039-492d-4bbb-85e6-87dbe94f364e service nova] Lock "df856146-61ba-4de7-a5fd-d75263927163-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.487902] env[63021]: DEBUG oslo_concurrency.lockutils [req-021358fd-6bee-4588-91ff-74d563232536 req-64356039-492d-4bbb-85e6-87dbe94f364e service nova] Lock "df856146-61ba-4de7-a5fd-d75263927163-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.488083] env[63021]: DEBUG nova.compute.manager [req-021358fd-6bee-4588-91ff-74d563232536 req-64356039-492d-4bbb-85e6-87dbe94f364e service nova] [instance: df856146-61ba-4de7-a5fd-d75263927163] No waiting events found dispatching network-vif-plugged-a43dcfd2-397e-44fe-bd1c-b6116c25a28a {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 851.488252] env[63021]: WARNING nova.compute.manager [req-021358fd-6bee-4588-91ff-74d563232536 req-64356039-492d-4bbb-85e6-87dbe94f364e service nova] [instance: df856146-61ba-4de7-a5fd-d75263927163] Received unexpected event network-vif-plugged-a43dcfd2-397e-44fe-bd1c-b6116c25a28a for instance with vm_state building and task_state spawning. [ 851.488414] env[63021]: DEBUG nova.compute.manager [req-021358fd-6bee-4588-91ff-74d563232536 req-64356039-492d-4bbb-85e6-87dbe94f364e service nova] [instance: df856146-61ba-4de7-a5fd-d75263927163] Received event network-changed-a43dcfd2-397e-44fe-bd1c-b6116c25a28a {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 851.488630] env[63021]: DEBUG nova.compute.manager [req-021358fd-6bee-4588-91ff-74d563232536 req-64356039-492d-4bbb-85e6-87dbe94f364e service nova] [instance: df856146-61ba-4de7-a5fd-d75263927163] Refreshing instance network info cache due to event network-changed-a43dcfd2-397e-44fe-bd1c-b6116c25a28a. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 851.488829] env[63021]: DEBUG oslo_concurrency.lockutils [req-021358fd-6bee-4588-91ff-74d563232536 req-64356039-492d-4bbb-85e6-87dbe94f364e service nova] Acquiring lock "refresh_cache-df856146-61ba-4de7-a5fd-d75263927163" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.491132] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e30ebd0-a532-4d7d-a51e-2d16d0341e88 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.499713] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0f4f03b-ac5b-4c31-ad49-1e0cae715010 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.531983] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c823b1a0-9374-41b4-8038-e25b95580583 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.540192] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ab73efe-9e3d-4dd6-830a-68f2390ab4c3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.556029] env[63021]: DEBUG nova.compute.provider_tree [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 851.697894] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293749, 'name': CreateVM_Task, 'duration_secs': 1.366067} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.699056] env[63021]: DEBUG nova.network.neutron [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 851.708032] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 851.708032] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.723049] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5285bae7-d36b-4e8e-a4e8-b07231e55670, 'name': SearchDatastore_Task, 'duration_secs': 0.030075} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.725532] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.725778] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 851.725995] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.726165] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.726331] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 851.726627] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.726922] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 851.727162] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-63658ec7-2736-4a27-a116-44d5bc0e2336 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.729314] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a3216f8d-5c5f-4b18-b332-841a21225bbe {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.735909] env[63021]: DEBUG oslo_vmware.api [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 851.735909] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52248794-6fda-8070-5af2-36cade458dd8" [ 851.735909] env[63021]: _type = "Task" [ 851.735909] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.740432] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 851.740625] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 851.741851] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ca39c28-8008-4af5-bacf-4fa400b13d3b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.747118] env[63021]: DEBUG oslo_vmware.api [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52248794-6fda-8070-5af2-36cade458dd8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.750684] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Waiting for the task: (returnval){ [ 851.750684] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5249166d-f78a-edba-f4cf-84554844ed0e" [ 851.750684] env[63021]: _type = "Task" [ 851.750684] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.772312] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5249166d-f78a-edba-f4cf-84554844ed0e, 'name': SearchDatastore_Task, 'duration_secs': 0.011678} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.772312] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e855b40-0800-4bd0-8610-a415d2fdf4a0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.777693] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Waiting for the task: (returnval){ [ 851.777693] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5240e6c4-c8da-4b7f-b08e-d2779f4c2495" [ 851.777693] env[63021]: _type = "Task" [ 851.777693] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.786297] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5240e6c4-c8da-4b7f-b08e-d2779f4c2495, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.852353] env[63021]: DEBUG oslo_vmware.api [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5215407b-195c-0045-0c3c-b4a147e486c6, 'name': SearchDatastore_Task, 'duration_secs': 0.048983} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.852776] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.853383] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 70f67905-7c60-433d-9ebc-d66fa44eb36e/70f67905-7c60-433d-9ebc-d66fa44eb36e.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 851.853784] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b401e5e9-3dae-4d89-9ada-d79a34a1f66e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.865327] env[63021]: DEBUG oslo_vmware.api [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Waiting for the task: (returnval){ [ 851.865327] env[63021]: value = "task-1293750" [ 851.865327] env[63021]: _type = "Task" [ 851.865327] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.874980] env[63021]: DEBUG oslo_vmware.api [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Task: {'id': task-1293750, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.877018] env[63021]: DEBUG nova.network.neutron [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Updating instance_info_cache with network_info: [{"id": "a43dcfd2-397e-44fe-bd1c-b6116c25a28a", "address": "fa:16:3e:c8:73:42", "network": {"id": "7258df3e-b1eb-4da6-ab52-5dadd2f87b51", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1084458843-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5661c44b534b4a7e9452e914fc80d66e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4f04accd-4a9f-41a1-b327-c96095e1ade4", "external-id": "nsx-vlan-transportzone-947", "segmentation_id": 947, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa43dcfd2-39", "ovs_interfaceid": "a43dcfd2-397e-44fe-bd1c-b6116c25a28a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.889592] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Releasing lock "refresh_cache-7787f6bb-e855-476f-9146-2a648932429d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.059516] env[63021]: DEBUG nova.scheduler.client.report [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 852.249121] env[63021]: DEBUG oslo_vmware.api [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52248794-6fda-8070-5af2-36cade458dd8, 'name': SearchDatastore_Task, 'duration_secs': 0.011111} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.250023] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.250023] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 852.250183] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.288606] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5240e6c4-c8da-4b7f-b08e-d2779f4c2495, 'name': SearchDatastore_Task, 'duration_secs': 0.010976} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.288882] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.289215] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 6f4e6245-c0d9-46b4-baf5-5aca19e66da5/6f4e6245-c0d9-46b4-baf5-5aca19e66da5.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 852.289524] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.289713] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 852.289939] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2bd83431-af5a-4f3f-9a0c-f06a98925a57 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.295024] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e9469b2b-a4c0-4989-9859-6bc6748865ce {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.304153] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Waiting for the task: (returnval){ [ 852.304153] env[63021]: value = "task-1293751" [ 852.304153] env[63021]: _type = "Task" [ 852.304153] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.310388] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 852.311115] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 852.312608] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-90507c0e-0e24-49c4-ac3a-fde4e3873f58 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.318567] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293751, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.326192] env[63021]: DEBUG oslo_vmware.api [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 852.326192] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5237cff9-dc75-1d14-0672-a61798cb0bed" [ 852.326192] env[63021]: _type = "Task" [ 852.326192] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.335795] env[63021]: DEBUG oslo_vmware.api [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5237cff9-dc75-1d14-0672-a61798cb0bed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.374675] env[63021]: DEBUG oslo_vmware.api [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Task: {'id': task-1293750, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.379414] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Releasing lock "refresh_cache-df856146-61ba-4de7-a5fd-d75263927163" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.379913] env[63021]: DEBUG nova.compute.manager [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Instance network_info: |[{"id": "a43dcfd2-397e-44fe-bd1c-b6116c25a28a", "address": "fa:16:3e:c8:73:42", "network": {"id": "7258df3e-b1eb-4da6-ab52-5dadd2f87b51", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1084458843-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5661c44b534b4a7e9452e914fc80d66e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4f04accd-4a9f-41a1-b327-c96095e1ade4", "external-id": "nsx-vlan-transportzone-947", "segmentation_id": 947, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa43dcfd2-39", "ovs_interfaceid": "a43dcfd2-397e-44fe-bd1c-b6116c25a28a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 852.380182] env[63021]: DEBUG oslo_concurrency.lockutils [req-021358fd-6bee-4588-91ff-74d563232536 req-64356039-492d-4bbb-85e6-87dbe94f364e service nova] Acquired lock "refresh_cache-df856146-61ba-4de7-a5fd-d75263927163" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.380365] env[63021]: DEBUG nova.network.neutron [req-021358fd-6bee-4588-91ff-74d563232536 req-64356039-492d-4bbb-85e6-87dbe94f364e service nova] [instance: df856146-61ba-4de7-a5fd-d75263927163] Refreshing network info cache for port a43dcfd2-397e-44fe-bd1c-b6116c25a28a {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 852.381615] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c8:73:42', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4f04accd-4a9f-41a1-b327-c96095e1ade4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a43dcfd2-397e-44fe-bd1c-b6116c25a28a', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 852.394335] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Creating folder: Project (5661c44b534b4a7e9452e914fc80d66e). Parent ref: group-v277447. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 852.396320] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0080ef14-8b49-488e-b98f-22388c5c5d86 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.414229] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Created folder: Project (5661c44b534b4a7e9452e914fc80d66e) in parent group-v277447. [ 852.414516] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Creating folder: Instances. Parent ref: group-v277510. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 852.415198] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0e0dbb35-0490-40ab-af84-63f4297c6301 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.421852] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 852.422238] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-480a3d61-43ff-4298-a22d-318a421b7f9b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.429798] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Created folder: Instances in parent group-v277510. [ 852.430058] env[63021]: DEBUG oslo.service.loopingcall [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 852.430262] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df856146-61ba-4de7-a5fd-d75263927163] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 852.434182] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4424c850-f8fe-46dc-ba6f-f03cdb2eb1cc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.447209] env[63021]: DEBUG oslo_vmware.api [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Waiting for the task: (returnval){ [ 852.447209] env[63021]: value = "task-1293754" [ 852.447209] env[63021]: _type = "Task" [ 852.447209] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.456498] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 852.456498] env[63021]: value = "task-1293755" [ 852.456498] env[63021]: _type = "Task" [ 852.456498] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.471725] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293755, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.566327] env[63021]: DEBUG oslo_concurrency.lockutils [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.458s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.566986] env[63021]: DEBUG nova.compute.manager [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 852.575021] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 23.172s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.575021] env[63021]: DEBUG nova.objects.instance [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63021) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 852.673294] env[63021]: DEBUG nova.network.neutron [req-021358fd-6bee-4588-91ff-74d563232536 req-64356039-492d-4bbb-85e6-87dbe94f364e service nova] [instance: df856146-61ba-4de7-a5fd-d75263927163] Updated VIF entry in instance network info cache for port a43dcfd2-397e-44fe-bd1c-b6116c25a28a. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 852.673826] env[63021]: DEBUG nova.network.neutron [req-021358fd-6bee-4588-91ff-74d563232536 req-64356039-492d-4bbb-85e6-87dbe94f364e service nova] [instance: df856146-61ba-4de7-a5fd-d75263927163] Updating instance_info_cache with network_info: [{"id": "a43dcfd2-397e-44fe-bd1c-b6116c25a28a", "address": "fa:16:3e:c8:73:42", "network": {"id": "7258df3e-b1eb-4da6-ab52-5dadd2f87b51", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1084458843-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5661c44b534b4a7e9452e914fc80d66e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4f04accd-4a9f-41a1-b327-c96095e1ade4", "external-id": "nsx-vlan-transportzone-947", "segmentation_id": 947, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa43dcfd2-39", "ovs_interfaceid": "a43dcfd2-397e-44fe-bd1c-b6116c25a28a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.815019] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293751, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.837241] env[63021]: DEBUG oslo_vmware.api [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5237cff9-dc75-1d14-0672-a61798cb0bed, 'name': SearchDatastore_Task, 'duration_secs': 0.023276} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.838521] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4cd03288-b507-4ce9-85a8-54521fc5a54b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.845190] env[63021]: DEBUG oslo_vmware.api [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 852.845190] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52f5b477-51d4-a917-cd4f-6037c6b2384d" [ 852.845190] env[63021]: _type = "Task" [ 852.845190] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.853767] env[63021]: DEBUG oslo_vmware.api [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52f5b477-51d4-a917-cd4f-6037c6b2384d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.874365] env[63021]: DEBUG oslo_vmware.api [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Task: {'id': task-1293750, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.632819} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.874686] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 70f67905-7c60-433d-9ebc-d66fa44eb36e/70f67905-7c60-433d-9ebc-d66fa44eb36e.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 852.875034] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 852.875220] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6a889f67-c7f8-4063-ba30-bf91665b4284 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.886821] env[63021]: DEBUG oslo_vmware.api [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Waiting for the task: (returnval){ [ 852.886821] env[63021]: value = "task-1293756" [ 852.886821] env[63021]: _type = "Task" [ 852.886821] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.893736] env[63021]: DEBUG oslo_vmware.api [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Task: {'id': task-1293756, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.967740] env[63021]: DEBUG oslo_vmware.api [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': task-1293754, 'name': PowerOffVM_Task, 'duration_secs': 0.26511} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.968444] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 852.969363] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e14a847-3f3d-44b5-a724-d84755b9f6fd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.975108] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293755, 'name': CreateVM_Task} progress is 25%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.993018] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0887f2a6-176d-4986-b27d-feac42731507 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.032702] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 853.033399] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-408d973b-f9f3-4b1d-8928-665c089f9a82 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.041917] env[63021]: DEBUG oslo_vmware.api [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Waiting for the task: (returnval){ [ 853.041917] env[63021]: value = "task-1293757" [ 853.041917] env[63021]: _type = "Task" [ 853.041917] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.053224] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] VM already powered off {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 853.053592] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 853.056487] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.056487] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.056487] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 853.056487] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e31e2130-2c25-46b6-b79f-b129c27c7d4f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.066020] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 853.066020] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 853.066020] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69e9519f-f32d-4aac-8856-6203c1b865a3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.071649] env[63021]: DEBUG oslo_vmware.api [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Waiting for the task: (returnval){ [ 853.071649] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52915c2d-347d-9f24-fe87-1dbd6f209ca5" [ 853.071649] env[63021]: _type = "Task" [ 853.071649] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.082756] env[63021]: DEBUG nova.compute.utils [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 853.084608] env[63021]: DEBUG oslo_vmware.api [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52915c2d-347d-9f24-fe87-1dbd6f209ca5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.085312] env[63021]: DEBUG nova.compute.manager [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 853.085603] env[63021]: DEBUG nova.network.neutron [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 853.176753] env[63021]: DEBUG oslo_concurrency.lockutils [req-021358fd-6bee-4588-91ff-74d563232536 req-64356039-492d-4bbb-85e6-87dbe94f364e service nova] Releasing lock "refresh_cache-df856146-61ba-4de7-a5fd-d75263927163" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.289761] env[63021]: DEBUG nova.policy [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cfc48d1f237f4597ad06f9d7a6bd8a00', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4b4c8cd929ff4068b45ee829b19f9769', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 853.316195] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293751, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.355140] env[63021]: DEBUG oslo_vmware.api [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52f5b477-51d4-a917-cd4f-6037c6b2384d, 'name': SearchDatastore_Task, 'duration_secs': 0.046048} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.355403] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.355652] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 71909213-258d-45f5-9a3e-e473e8ce8aa2/71909213-258d-45f5-9a3e-e473e8ce8aa2.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 853.355897] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-728691bc-a4ac-458f-b0a1-e1a90d0f4684 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.363345] env[63021]: DEBUG oslo_vmware.api [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 853.363345] env[63021]: value = "task-1293758" [ 853.363345] env[63021]: _type = "Task" [ 853.363345] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.371601] env[63021]: DEBUG oslo_vmware.api [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1293758, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.393366] env[63021]: DEBUG oslo_vmware.api [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Task: {'id': task-1293756, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.114169} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.393647] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 853.394524] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb7bc7d2-97e1-4702-998a-57c31dcbd77c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.417511] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] 70f67905-7c60-433d-9ebc-d66fa44eb36e/70f67905-7c60-433d-9ebc-d66fa44eb36e.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 853.417835] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e29908a0-5b4c-48e6-a97d-42a16a004d9d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.437682] env[63021]: DEBUG oslo_vmware.api [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Waiting for the task: (returnval){ [ 853.437682] env[63021]: value = "task-1293759" [ 853.437682] env[63021]: _type = "Task" [ 853.437682] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.447663] env[63021]: DEBUG oslo_vmware.api [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Task: {'id': task-1293759, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.470110] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293755, 'name': CreateVM_Task, 'duration_secs': 0.70534} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.470285] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df856146-61ba-4de7-a5fd-d75263927163] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 853.470987] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.471179] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.471513] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 853.471782] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-989fd5f7-33d3-4ef4-8665-a13d87823fec {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.479758] env[63021]: DEBUG oslo_vmware.api [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 853.479758] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52a5d1de-eab6-e74d-b702-0af96cd2dd9e" [ 853.479758] env[63021]: _type = "Task" [ 853.479758] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.498449] env[63021]: DEBUG oslo_vmware.api [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52a5d1de-eab6-e74d-b702-0af96cd2dd9e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.562314] env[63021]: DEBUG nova.network.neutron [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Successfully created port: 43e48c53-d06d-41ff-97dd-a7747d5551d9 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 853.584331] env[63021]: DEBUG oslo_vmware.api [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52915c2d-347d-9f24-fe87-1dbd6f209ca5, 'name': SearchDatastore_Task, 'duration_secs': 0.009782} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.584331] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-46d2e101-7575-42e0-b503-a9407688fb76 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.588020] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d496930b-0c48-4397-99da-df584a673c40 tempest-ServersAdmin275Test-1323225566 tempest-ServersAdmin275Test-1323225566-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.588459] env[63021]: DEBUG nova.compute.manager [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 853.591196] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.103s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.591400] env[63021]: DEBUG nova.objects.instance [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Lazy-loading 'resources' on Instance uuid 868b1aeb-8109-4cb9-bb58-33f3d92989ad {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 853.595717] env[63021]: DEBUG oslo_vmware.api [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Waiting for the task: (returnval){ [ 853.595717] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52ba84df-b523-2609-69b1-d8394fffaa6b" [ 853.595717] env[63021]: _type = "Task" [ 853.595717] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.605266] env[63021]: DEBUG oslo_vmware.api [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52ba84df-b523-2609-69b1-d8394fffaa6b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.815078] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293751, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.873897] env[63021]: DEBUG oslo_vmware.api [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1293758, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.949617] env[63021]: DEBUG oslo_vmware.api [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Task: {'id': task-1293759, 'name': ReconfigVM_Task, 'duration_secs': 0.267229} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.949903] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Reconfigured VM instance instance-00000040 to attach disk [datastore1] 70f67905-7c60-433d-9ebc-d66fa44eb36e/70f67905-7c60-433d-9ebc-d66fa44eb36e.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 853.950604] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2e05e44c-d48c-4c92-b432-b374ee9e61aa {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.959962] env[63021]: DEBUG oslo_vmware.api [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Waiting for the task: (returnval){ [ 853.959962] env[63021]: value = "task-1293760" [ 853.959962] env[63021]: _type = "Task" [ 853.959962] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.970031] env[63021]: DEBUG oslo_vmware.api [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Task: {'id': task-1293760, 'name': Rename_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.990266] env[63021]: DEBUG oslo_vmware.api [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52a5d1de-eab6-e74d-b702-0af96cd2dd9e, 'name': SearchDatastore_Task, 'duration_secs': 0.02553} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.990618] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.990885] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 853.991171] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.991334] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.991549] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 853.991865] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a5390cd4-e373-4f4a-9025-99d8702d04f5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.001457] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 854.001651] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 854.002442] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-15f3e640-d659-4cb0-85d8-e8ef910e35d3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.009181] env[63021]: DEBUG oslo_vmware.api [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 854.009181] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5286d3b0-1d99-287e-4178-c36274962ada" [ 854.009181] env[63021]: _type = "Task" [ 854.009181] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.017327] env[63021]: DEBUG oslo_vmware.api [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5286d3b0-1d99-287e-4178-c36274962ada, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.112461] env[63021]: DEBUG oslo_vmware.api [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52ba84df-b523-2609-69b1-d8394fffaa6b, 'name': SearchDatastore_Task, 'duration_secs': 0.011058} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.112793] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.113476] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 7787f6bb-e855-476f-9146-2a648932429d/e624225d-af0f-415c-90cc-d4b228c479f6-rescue.vmdk. {{(pid=63021) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 854.113476] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b1e5eddd-b9f8-409e-848c-a91e5cc372e0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.123459] env[63021]: DEBUG oslo_vmware.api [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Waiting for the task: (returnval){ [ 854.123459] env[63021]: value = "task-1293761" [ 854.123459] env[63021]: _type = "Task" [ 854.123459] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.136796] env[63021]: DEBUG oslo_vmware.api [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': task-1293761, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.315953] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293751, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.608642} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.318963] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 6f4e6245-c0d9-46b4-baf5-5aca19e66da5/6f4e6245-c0d9-46b4-baf5-5aca19e66da5.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 854.319269] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 854.319774] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1843fe84-2a4b-4c39-90ee-82a0853650db {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.330346] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Waiting for the task: (returnval){ [ 854.330346] env[63021]: value = "task-1293762" [ 854.330346] env[63021]: _type = "Task" [ 854.330346] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.342749] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293762, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.378158] env[63021]: DEBUG oslo_vmware.api [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1293758, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.013557} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.378374] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 71909213-258d-45f5-9a3e-e473e8ce8aa2/71909213-258d-45f5-9a3e-e473e8ce8aa2.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 854.378642] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 854.378946] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5518ac60-6e26-4552-b973-32e02d4c63ad {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.389069] env[63021]: DEBUG oslo_vmware.api [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 854.389069] env[63021]: value = "task-1293763" [ 854.389069] env[63021]: _type = "Task" [ 854.389069] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.399683] env[63021]: DEBUG oslo_vmware.api [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1293763, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.442146] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ae97b4f-2915-40ea-85e4-203712f0d157 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.450273] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-377fa1b3-7145-453a-867d-69f4a525b674 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.485193] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b90310a-09fe-4ce9-9668-66051780ae86 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.501074] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de88258c-5875-4b5e-b0d7-013176b9b905 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.505636] env[63021]: DEBUG oslo_vmware.api [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Task: {'id': task-1293760, 'name': Rename_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.517352] env[63021]: DEBUG nova.compute.provider_tree [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 854.528500] env[63021]: DEBUG oslo_vmware.api [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5286d3b0-1d99-287e-4178-c36274962ada, 'name': SearchDatastore_Task, 'duration_secs': 0.03049} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.529968] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9213c9e7-fc16-4e0a-ad4c-b7456524cafb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.537170] env[63021]: DEBUG oslo_vmware.api [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 854.537170] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52ad2aae-7746-1bff-3af6-a97b453ae181" [ 854.537170] env[63021]: _type = "Task" [ 854.537170] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.546476] env[63021]: DEBUG oslo_vmware.api [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52ad2aae-7746-1bff-3af6-a97b453ae181, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.603697] env[63021]: DEBUG nova.compute.manager [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 854.633599] env[63021]: DEBUG nova.virt.hardware [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 854.633867] env[63021]: DEBUG nova.virt.hardware [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 854.634034] env[63021]: DEBUG nova.virt.hardware [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 854.634290] env[63021]: DEBUG nova.virt.hardware [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 854.634447] env[63021]: DEBUG nova.virt.hardware [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 854.634592] env[63021]: DEBUG nova.virt.hardware [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 854.634819] env[63021]: DEBUG nova.virt.hardware [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 854.634992] env[63021]: DEBUG nova.virt.hardware [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 854.635175] env[63021]: DEBUG nova.virt.hardware [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 854.635360] env[63021]: DEBUG nova.virt.hardware [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 854.635541] env[63021]: DEBUG nova.virt.hardware [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 854.636398] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35da622c-a5d4-4a71-a1a9-39f2e5322fc6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.642393] env[63021]: DEBUG oslo_vmware.api [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': task-1293761, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.648901] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6af8e122-09d6-4e46-881e-0d5f4c267927 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.843099] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293762, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073536} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.843379] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 854.844212] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-934a93a5-bb20-45ec-8a01-969f08a9484d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.864370] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 6f4e6245-c0d9-46b4-baf5-5aca19e66da5/6f4e6245-c0d9-46b4-baf5-5aca19e66da5.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 854.864908] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-14d96ae7-7692-495d-b52f-2ee24809bc40 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.884563] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Waiting for the task: (returnval){ [ 854.884563] env[63021]: value = "task-1293764" [ 854.884563] env[63021]: _type = "Task" [ 854.884563] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.893230] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293764, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.901104] env[63021]: DEBUG oslo_vmware.api [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1293763, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065374} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.901374] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 854.902129] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-441859fd-666e-4907-975f-b5b51f413926 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.925586] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] 71909213-258d-45f5-9a3e-e473e8ce8aa2/71909213-258d-45f5-9a3e-e473e8ce8aa2.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 854.925719] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-69968554-d723-498e-90c3-b493a950d916 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.946583] env[63021]: DEBUG oslo_vmware.api [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 854.946583] env[63021]: value = "task-1293765" [ 854.946583] env[63021]: _type = "Task" [ 854.946583] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.955472] env[63021]: DEBUG oslo_vmware.api [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1293765, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.992060] env[63021]: DEBUG oslo_vmware.api [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Task: {'id': task-1293760, 'name': Rename_Task, 'duration_secs': 0.831945} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.992664] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 854.992664] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f548b289-44ba-4f59-bdad-04096c2afb3f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.001111] env[63021]: DEBUG oslo_vmware.api [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Waiting for the task: (returnval){ [ 855.001111] env[63021]: value = "task-1293766" [ 855.001111] env[63021]: _type = "Task" [ 855.001111] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.011060] env[63021]: DEBUG oslo_vmware.api [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Task: {'id': task-1293766, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.024301] env[63021]: DEBUG nova.scheduler.client.report [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 855.048203] env[63021]: DEBUG oslo_vmware.api [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52ad2aae-7746-1bff-3af6-a97b453ae181, 'name': SearchDatastore_Task, 'duration_secs': 0.017683} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.048517] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.048783] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] df856146-61ba-4de7-a5fd-d75263927163/df856146-61ba-4de7-a5fd-d75263927163.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 855.049210] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7b244d91-d068-4a8e-af81-c75bad730d56 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.058848] env[63021]: DEBUG oslo_vmware.api [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 855.058848] env[63021]: value = "task-1293767" [ 855.058848] env[63021]: _type = "Task" [ 855.058848] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.069961] env[63021]: DEBUG oslo_vmware.api [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1293767, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.139159] env[63021]: DEBUG oslo_vmware.api [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': task-1293761, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.67154} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.139474] env[63021]: INFO nova.virt.vmwareapi.ds_util [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 7787f6bb-e855-476f-9146-2a648932429d/e624225d-af0f-415c-90cc-d4b228c479f6-rescue.vmdk. [ 855.140343] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-579ae96a-2302-4627-a1c8-4f4d8cc31b8c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.169960] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] 7787f6bb-e855-476f-9146-2a648932429d/e624225d-af0f-415c-90cc-d4b228c479f6-rescue.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 855.171484] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-13537bc7-b794-4bf5-b068-880ce61a0546 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.186013] env[63021]: DEBUG nova.compute.manager [req-c9fcdd06-79a6-4dec-97b4-d5f12aac14bb req-e56fc13a-f6a7-46fa-a118-c6673acd774f service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Received event network-vif-plugged-43e48c53-d06d-41ff-97dd-a7747d5551d9 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 855.186234] env[63021]: DEBUG oslo_concurrency.lockutils [req-c9fcdd06-79a6-4dec-97b4-d5f12aac14bb req-e56fc13a-f6a7-46fa-a118-c6673acd774f service nova] Acquiring lock "45c9e2be-eb19-4b83-b280-c9eeaddfccba-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.186462] env[63021]: DEBUG oslo_concurrency.lockutils [req-c9fcdd06-79a6-4dec-97b4-d5f12aac14bb req-e56fc13a-f6a7-46fa-a118-c6673acd774f service nova] Lock "45c9e2be-eb19-4b83-b280-c9eeaddfccba-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.186630] env[63021]: DEBUG oslo_concurrency.lockutils [req-c9fcdd06-79a6-4dec-97b4-d5f12aac14bb req-e56fc13a-f6a7-46fa-a118-c6673acd774f service nova] Lock "45c9e2be-eb19-4b83-b280-c9eeaddfccba-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.186792] env[63021]: DEBUG nova.compute.manager [req-c9fcdd06-79a6-4dec-97b4-d5f12aac14bb req-e56fc13a-f6a7-46fa-a118-c6673acd774f service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] No waiting events found dispatching network-vif-plugged-43e48c53-d06d-41ff-97dd-a7747d5551d9 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 855.186970] env[63021]: WARNING nova.compute.manager [req-c9fcdd06-79a6-4dec-97b4-d5f12aac14bb req-e56fc13a-f6a7-46fa-a118-c6673acd774f service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Received unexpected event network-vif-plugged-43e48c53-d06d-41ff-97dd-a7747d5551d9 for instance with vm_state building and task_state spawning. [ 855.195062] env[63021]: DEBUG oslo_vmware.api [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Waiting for the task: (returnval){ [ 855.195062] env[63021]: value = "task-1293768" [ 855.195062] env[63021]: _type = "Task" [ 855.195062] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.204279] env[63021]: DEBUG oslo_vmware.api [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': task-1293768, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.396323] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293764, 'name': ReconfigVM_Task, 'duration_secs': 0.29933} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.396678] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 6f4e6245-c0d9-46b4-baf5-5aca19e66da5/6f4e6245-c0d9-46b4-baf5-5aca19e66da5.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 855.397588] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-74815d64-596d-4753-b5af-1261acb38cc9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.405636] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Waiting for the task: (returnval){ [ 855.405636] env[63021]: value = "task-1293769" [ 855.405636] env[63021]: _type = "Task" [ 855.405636] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.414639] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293769, 'name': Rename_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.457493] env[63021]: DEBUG oslo_vmware.api [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1293765, 'name': ReconfigVM_Task, 'duration_secs': 0.383444} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.457759] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Reconfigured VM instance instance-00000042 to attach disk [datastore2] 71909213-258d-45f5-9a3e-e473e8ce8aa2/71909213-258d-45f5-9a3e-e473e8ce8aa2.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 855.458431] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b6dc09f1-54c5-45a3-85ba-5d57527511f1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.467151] env[63021]: DEBUG oslo_vmware.api [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 855.467151] env[63021]: value = "task-1293770" [ 855.467151] env[63021]: _type = "Task" [ 855.467151] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.476896] env[63021]: DEBUG oslo_vmware.api [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1293770, 'name': Rename_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.512201] env[63021]: DEBUG oslo_vmware.api [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Task: {'id': task-1293766, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.532038] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.941s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.539501] env[63021]: DEBUG nova.network.neutron [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Successfully updated port: 43e48c53-d06d-41ff-97dd-a7747d5551d9 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 855.540908] env[63021]: DEBUG oslo_concurrency.lockutils [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.470s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.545098] env[63021]: INFO nova.compute.claims [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 855.565115] env[63021]: INFO nova.scheduler.client.report [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Deleted allocations for instance 868b1aeb-8109-4cb9-bb58-33f3d92989ad [ 855.576464] env[63021]: DEBUG oslo_vmware.api [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1293767, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.506406} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.576689] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] df856146-61ba-4de7-a5fd-d75263927163/df856146-61ba-4de7-a5fd-d75263927163.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 855.576957] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 855.577243] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-33d7cdf5-b343-48d0-94f3-25b1758e6446 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.585061] env[63021]: DEBUG oslo_vmware.api [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 855.585061] env[63021]: value = "task-1293771" [ 855.585061] env[63021]: _type = "Task" [ 855.585061] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.595999] env[63021]: DEBUG oslo_vmware.api [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1293771, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.630019] env[63021]: DEBUG nova.compute.manager [req-ccdfc546-5933-4454-98bf-1b83d698ee7e req-115f9f99-0170-4c90-9bd1-c941a8a08bf6 service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Received event network-changed-43e48c53-d06d-41ff-97dd-a7747d5551d9 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 855.630756] env[63021]: DEBUG nova.compute.manager [req-ccdfc546-5933-4454-98bf-1b83d698ee7e req-115f9f99-0170-4c90-9bd1-c941a8a08bf6 service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Refreshing instance network info cache due to event network-changed-43e48c53-d06d-41ff-97dd-a7747d5551d9. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 855.630756] env[63021]: DEBUG oslo_concurrency.lockutils [req-ccdfc546-5933-4454-98bf-1b83d698ee7e req-115f9f99-0170-4c90-9bd1-c941a8a08bf6 service nova] Acquiring lock "refresh_cache-45c9e2be-eb19-4b83-b280-c9eeaddfccba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.630756] env[63021]: DEBUG oslo_concurrency.lockutils [req-ccdfc546-5933-4454-98bf-1b83d698ee7e req-115f9f99-0170-4c90-9bd1-c941a8a08bf6 service nova] Acquired lock "refresh_cache-45c9e2be-eb19-4b83-b280-c9eeaddfccba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.630756] env[63021]: DEBUG nova.network.neutron [req-ccdfc546-5933-4454-98bf-1b83d698ee7e req-115f9f99-0170-4c90-9bd1-c941a8a08bf6 service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Refreshing network info cache for port 43e48c53-d06d-41ff-97dd-a7747d5551d9 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 855.705738] env[63021]: DEBUG oslo_vmware.api [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': task-1293768, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.916045] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293769, 'name': Rename_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.977708] env[63021]: DEBUG oslo_vmware.api [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1293770, 'name': Rename_Task, 'duration_secs': 0.339774} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.978015] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 855.978296] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b8d07f0c-016c-488f-9d2b-25422f114080 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.987763] env[63021]: DEBUG oslo_vmware.api [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 855.987763] env[63021]: value = "task-1293772" [ 855.987763] env[63021]: _type = "Task" [ 855.987763] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.996016] env[63021]: DEBUG oslo_vmware.api [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1293772, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.012046] env[63021]: DEBUG oslo_vmware.api [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Task: {'id': task-1293766, 'name': PowerOnVM_Task, 'duration_secs': 0.742565} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.012333] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 856.012541] env[63021]: INFO nova.compute.manager [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Took 9.81 seconds to spawn the instance on the hypervisor. [ 856.012714] env[63021]: DEBUG nova.compute.manager [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 856.013474] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43a29eaf-fd0b-4c04-af84-ce8fd843f53f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.050076] env[63021]: DEBUG oslo_concurrency.lockutils [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Acquiring lock "refresh_cache-45c9e2be-eb19-4b83-b280-c9eeaddfccba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.073580] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0b87aeb7-89d2-4a71-9efb-277b660e1fa9 tempest-VolumesAdminNegativeTest-1107591338 tempest-VolumesAdminNegativeTest-1107591338-project-member] Lock "868b1aeb-8109-4cb9-bb58-33f3d92989ad" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.471s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.096451] env[63021]: DEBUG oslo_vmware.api [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1293771, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061625} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.096749] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 856.097568] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07d88bc9-dedb-4982-9b43-f6278f0c4c6f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.122936] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] df856146-61ba-4de7-a5fd-d75263927163/df856146-61ba-4de7-a5fd-d75263927163.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 856.123299] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b0bf9c17-bdfa-4089-8f0b-6fd0d367d0bc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.148141] env[63021]: DEBUG oslo_vmware.api [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 856.148141] env[63021]: value = "task-1293773" [ 856.148141] env[63021]: _type = "Task" [ 856.148141] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.160085] env[63021]: DEBUG oslo_vmware.api [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1293773, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.184172] env[63021]: DEBUG nova.network.neutron [req-ccdfc546-5933-4454-98bf-1b83d698ee7e req-115f9f99-0170-4c90-9bd1-c941a8a08bf6 service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 856.210678] env[63021]: DEBUG oslo_vmware.api [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': task-1293768, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.287110] env[63021]: DEBUG nova.network.neutron [req-ccdfc546-5933-4454-98bf-1b83d698ee7e req-115f9f99-0170-4c90-9bd1-c941a8a08bf6 service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.417656] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293769, 'name': Rename_Task, 'duration_secs': 0.876192} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.417868] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 856.418150] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-af268efc-a2a2-499f-b23a-c2799efc1184 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.426085] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Waiting for the task: (returnval){ [ 856.426085] env[63021]: value = "task-1293774" [ 856.426085] env[63021]: _type = "Task" [ 856.426085] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.435962] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293774, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.500343] env[63021]: DEBUG oslo_vmware.api [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1293772, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.530839] env[63021]: INFO nova.compute.manager [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Took 37.44 seconds to build instance. [ 856.661928] env[63021]: DEBUG oslo_vmware.api [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1293773, 'name': ReconfigVM_Task, 'duration_secs': 0.316433} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.662419] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Reconfigured VM instance instance-00000043 to attach disk [datastore2] df856146-61ba-4de7-a5fd-d75263927163/df856146-61ba-4de7-a5fd-d75263927163.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 856.663209] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-51809bed-778d-4753-9582-7d5b24d6d431 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.671670] env[63021]: DEBUG oslo_vmware.api [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 856.671670] env[63021]: value = "task-1293775" [ 856.671670] env[63021]: _type = "Task" [ 856.671670] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.681096] env[63021]: DEBUG oslo_vmware.api [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1293775, 'name': Rename_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.711747] env[63021]: DEBUG oslo_vmware.api [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': task-1293768, 'name': ReconfigVM_Task, 'duration_secs': 1.300379} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.715923] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Reconfigured VM instance instance-00000041 to attach disk [datastore1] 7787f6bb-e855-476f-9146-2a648932429d/e624225d-af0f-415c-90cc-d4b228c479f6-rescue.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 856.719186] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22d4243b-7add-4fc1-97dc-1aacfa3bd72d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.750257] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-81c5f79a-72f3-4e37-860c-b3eb8f044233 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.767886] env[63021]: DEBUG oslo_vmware.api [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Waiting for the task: (returnval){ [ 856.767886] env[63021]: value = "task-1293776" [ 856.767886] env[63021]: _type = "Task" [ 856.767886] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.777914] env[63021]: DEBUG oslo_vmware.api [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': task-1293776, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.791193] env[63021]: DEBUG oslo_concurrency.lockutils [req-ccdfc546-5933-4454-98bf-1b83d698ee7e req-115f9f99-0170-4c90-9bd1-c941a8a08bf6 service nova] Releasing lock "refresh_cache-45c9e2be-eb19-4b83-b280-c9eeaddfccba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.791193] env[63021]: DEBUG oslo_concurrency.lockutils [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Acquired lock "refresh_cache-45c9e2be-eb19-4b83-b280-c9eeaddfccba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.791193] env[63021]: DEBUG nova.network.neutron [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 856.940857] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293774, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.946491] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9972ae51-f38f-4c9a-98e8-116a3b5d1549 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.957018] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41f02f2a-5a29-451a-a9d2-b149b1befb04 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.986162] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d10031d2-9577-4f55-a247-beb549cb51cd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.001756] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e77a6e51-974f-4334-9293-1606f3acaefe {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.006311] env[63021]: DEBUG oslo_vmware.api [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1293772, 'name': PowerOnVM_Task, 'duration_secs': 0.939351} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.006607] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 857.006831] env[63021]: INFO nova.compute.manager [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Took 9.35 seconds to spawn the instance on the hypervisor. [ 857.007020] env[63021]: DEBUG nova.compute.manager [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 857.008212] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aad4b1c-3629-4ad8-881d-e6ac5138c3a5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.019671] env[63021]: DEBUG nova.compute.provider_tree [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 857.032494] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f150420b-4150-4415-bc24-1a64372b6d0d tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Lock "70f67905-7c60-433d-9ebc-d66fa44eb36e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.715s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.182689] env[63021]: DEBUG oslo_vmware.api [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1293775, 'name': Rename_Task, 'duration_secs': 0.183704} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.182932] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 857.183191] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e619851f-d8f1-4b08-958d-e2994567ede6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.190366] env[63021]: DEBUG oslo_vmware.api [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 857.190366] env[63021]: value = "task-1293777" [ 857.190366] env[63021]: _type = "Task" [ 857.190366] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.198322] env[63021]: DEBUG oslo_vmware.api [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1293777, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.281970] env[63021]: DEBUG oslo_vmware.api [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': task-1293776, 'name': ReconfigVM_Task, 'duration_secs': 0.209316} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.282363] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 857.282796] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7aae0cad-4fc9-4565-8513-f8eff32b63cf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.290766] env[63021]: DEBUG oslo_vmware.api [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Waiting for the task: (returnval){ [ 857.290766] env[63021]: value = "task-1293778" [ 857.290766] env[63021]: _type = "Task" [ 857.290766] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.300456] env[63021]: DEBUG oslo_vmware.api [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': task-1293778, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.330695] env[63021]: DEBUG nova.network.neutron [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 857.439063] env[63021]: DEBUG oslo_vmware.api [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293774, 'name': PowerOnVM_Task, 'duration_secs': 0.527967} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.439063] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 857.443578] env[63021]: DEBUG nova.compute.manager [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 857.444296] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-530f8954-3a74-43c8-943f-3407e5efb3da {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.519705] env[63021]: DEBUG nova.network.neutron [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Updating instance_info_cache with network_info: [{"id": "43e48c53-d06d-41ff-97dd-a7747d5551d9", "address": "fa:16:3e:73:e1:9e", "network": {"id": "36548b96-bd4d-4725-8dab-5cb1cca12c4e", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-672805719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b4c8cd929ff4068b45ee829b19f9769", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8614db14-cc04-466b-b309-367ab8296cda", "external-id": "nsx-vlan-transportzone-337", "segmentation_id": 337, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43e48c53-d0", "ovs_interfaceid": "43e48c53-d06d-41ff-97dd-a7747d5551d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.523430] env[63021]: DEBUG nova.scheduler.client.report [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 857.535900] env[63021]: DEBUG nova.compute.manager [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 857.540356] env[63021]: INFO nova.compute.manager [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Took 37.10 seconds to build instance. [ 857.701874] env[63021]: DEBUG oslo_vmware.api [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1293777, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.803706] env[63021]: DEBUG oslo_vmware.api [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': task-1293778, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.968043] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.027031] env[63021]: DEBUG oslo_concurrency.lockutils [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Releasing lock "refresh_cache-45c9e2be-eb19-4b83-b280-c9eeaddfccba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.027031] env[63021]: DEBUG nova.compute.manager [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Instance network_info: |[{"id": "43e48c53-d06d-41ff-97dd-a7747d5551d9", "address": "fa:16:3e:73:e1:9e", "network": {"id": "36548b96-bd4d-4725-8dab-5cb1cca12c4e", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-672805719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b4c8cd929ff4068b45ee829b19f9769", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8614db14-cc04-466b-b309-367ab8296cda", "external-id": "nsx-vlan-transportzone-337", "segmentation_id": 337, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43e48c53-d0", "ovs_interfaceid": "43e48c53-d06d-41ff-97dd-a7747d5551d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 858.027276] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:73:e1:9e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8614db14-cc04-466b-b309-367ab8296cda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '43e48c53-d06d-41ff-97dd-a7747d5551d9', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 858.033738] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Creating folder: Project (4b4c8cd929ff4068b45ee829b19f9769). Parent ref: group-v277447. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 858.039254] env[63021]: DEBUG oslo_concurrency.lockutils [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.497s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.039254] env[63021]: DEBUG nova.compute.manager [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 858.041687] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d1f256a6-e93b-4ce2-ab55-943e8f38b2ed {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.043868] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.470s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.045417] env[63021]: INFO nova.compute.claims [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 858.049971] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1315b751-eb1c-4c42-bffc-01e8fd3321e5 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "71909213-258d-45f5-9a3e-e473e8ce8aa2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.843s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.066639] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Created folder: Project (4b4c8cd929ff4068b45ee829b19f9769) in parent group-v277447. [ 858.066878] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Creating folder: Instances. Parent ref: group-v277513. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 858.067472] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-54feb963-f2e7-45ef-b2c3-f191c0aaea81 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.074881] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.082807] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Created folder: Instances in parent group-v277513. [ 858.083103] env[63021]: DEBUG oslo.service.loopingcall [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 858.083300] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 858.083514] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c9612684-0517-4da6-9941-e85ce7691b01 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.108025] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 858.108025] env[63021]: value = "task-1293781" [ 858.108025] env[63021]: _type = "Task" [ 858.108025] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.118896] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Acquiring lock "70f67905-7c60-433d-9ebc-d66fa44eb36e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.119198] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Lock "70f67905-7c60-433d-9ebc-d66fa44eb36e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.119396] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Acquiring lock "70f67905-7c60-433d-9ebc-d66fa44eb36e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.119569] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Lock "70f67905-7c60-433d-9ebc-d66fa44eb36e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.119727] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Lock "70f67905-7c60-433d-9ebc-d66fa44eb36e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.126299] env[63021]: INFO nova.compute.manager [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Terminating instance [ 858.126299] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293781, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.126299] env[63021]: DEBUG nova.compute.manager [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 858.126299] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 858.126716] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90afe132-1bb1-4393-a732-2e35a9b4a816 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.134818] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 858.135054] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6e196dff-5784-4585-aa3b-a92e0447ad50 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.142424] env[63021]: DEBUG oslo_vmware.api [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Waiting for the task: (returnval){ [ 858.142424] env[63021]: value = "task-1293782" [ 858.142424] env[63021]: _type = "Task" [ 858.142424] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.151944] env[63021]: DEBUG oslo_vmware.api [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Task: {'id': task-1293782, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.206599] env[63021]: DEBUG oslo_vmware.api [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1293777, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.302080] env[63021]: DEBUG oslo_vmware.api [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': task-1293778, 'name': PowerOnVM_Task, 'duration_secs': 1.010693} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.302399] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 858.305496] env[63021]: DEBUG nova.compute.manager [None req-e5ce4d0a-c5c1-4f11-a588-ca696c5827fa tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 858.306371] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e753b3ef-116b-40d2-9f63-80d4c2f9da2b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.387701] env[63021]: DEBUG oslo_concurrency.lockutils [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Acquiring lock "6f4e6245-c0d9-46b4-baf5-5aca19e66da5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.387815] env[63021]: DEBUG oslo_concurrency.lockutils [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Lock "6f4e6245-c0d9-46b4-baf5-5aca19e66da5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.388059] env[63021]: DEBUG oslo_concurrency.lockutils [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Acquiring lock "6f4e6245-c0d9-46b4-baf5-5aca19e66da5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.388273] env[63021]: DEBUG oslo_concurrency.lockutils [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Lock "6f4e6245-c0d9-46b4-baf5-5aca19e66da5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.388446] env[63021]: DEBUG oslo_concurrency.lockutils [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Lock "6f4e6245-c0d9-46b4-baf5-5aca19e66da5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.390824] env[63021]: INFO nova.compute.manager [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Terminating instance [ 858.392710] env[63021]: DEBUG oslo_concurrency.lockutils [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Acquiring lock "refresh_cache-6f4e6245-c0d9-46b4-baf5-5aca19e66da5" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.392947] env[63021]: DEBUG oslo_concurrency.lockutils [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Acquired lock "refresh_cache-6f4e6245-c0d9-46b4-baf5-5aca19e66da5" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.393211] env[63021]: DEBUG nova.network.neutron [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 858.545420] env[63021]: DEBUG nova.compute.utils [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 858.546842] env[63021]: DEBUG nova.compute.manager [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 858.547442] env[63021]: DEBUG nova.network.neutron [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 858.556026] env[63021]: DEBUG nova.compute.manager [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 858.600534] env[63021]: DEBUG nova.policy [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '31c5d292015043aa8bfc6647e7405ea3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5af87713ae6b477ca8a584aed287889d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 858.621160] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293781, 'name': CreateVM_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.656347] env[63021]: DEBUG oslo_vmware.api [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Task: {'id': task-1293782, 'name': PowerOffVM_Task, 'duration_secs': 0.32875} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.656693] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 858.656873] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 858.657148] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fa997b5d-fcd6-47ff-bcf5-aaf079ac7a89 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.701761] env[63021]: DEBUG oslo_vmware.api [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1293777, 'name': PowerOnVM_Task, 'duration_secs': 1.147643} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.702093] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 858.702311] env[63021]: INFO nova.compute.manager [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Took 8.51 seconds to spawn the instance on the hypervisor. [ 858.702489] env[63021]: DEBUG nova.compute.manager [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 858.703308] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d60deb7-6f62-4ff0-a579-df1488b59467 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.742128] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 858.742576] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 858.742576] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Deleting the datastore file [datastore1] 70f67905-7c60-433d-9ebc-d66fa44eb36e {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 858.743402] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-92e7746f-6359-400e-89ff-5900d7df4284 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.750177] env[63021]: DEBUG oslo_vmware.api [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Waiting for the task: (returnval){ [ 858.750177] env[63021]: value = "task-1293784" [ 858.750177] env[63021]: _type = "Task" [ 858.750177] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.759396] env[63021]: DEBUG oslo_vmware.api [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Task: {'id': task-1293784, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.918450] env[63021]: DEBUG nova.network.neutron [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 858.920788] env[63021]: DEBUG nova.network.neutron [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Successfully created port: eff6f35a-0a13-4e0c-88b0-96b66b240f9b {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 859.011427] env[63021]: DEBUG nova.network.neutron [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.051394] env[63021]: DEBUG nova.compute.manager [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 859.088346] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.125499] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293781, 'name': CreateVM_Task, 'duration_secs': 0.548113} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.125945] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 859.126655] env[63021]: DEBUG oslo_concurrency.lockutils [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.126814] env[63021]: DEBUG oslo_concurrency.lockutils [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.127139] env[63021]: DEBUG oslo_concurrency.lockutils [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 859.127402] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b21b45ab-ee45-4c97-8f1c-558507d4321a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.135478] env[63021]: DEBUG oslo_vmware.api [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Waiting for the task: (returnval){ [ 859.135478] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]522dacaa-3b72-b2c8-725d-70a9f4b1ada4" [ 859.135478] env[63021]: _type = "Task" [ 859.135478] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.146857] env[63021]: DEBUG oslo_vmware.api [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]522dacaa-3b72-b2c8-725d-70a9f4b1ada4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.227325] env[63021]: INFO nova.compute.manager [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Took 36.75 seconds to build instance. [ 859.263305] env[63021]: DEBUG oslo_vmware.api [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Task: {'id': task-1293784, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.260882} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.263575] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 859.264159] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 859.264407] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 859.264589] env[63021]: INFO nova.compute.manager [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Took 1.14 seconds to destroy the instance on the hypervisor. [ 859.264832] env[63021]: DEBUG oslo.service.loopingcall [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 859.265031] env[63021]: DEBUG nova.compute.manager [-] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 859.265128] env[63021]: DEBUG nova.network.neutron [-] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 859.406740] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a78c7c5b-2eaf-494e-a857-585486733e6d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.416051] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a657d071-7054-462c-9ab0-59e8408dc7a0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.448655] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd8a2544-0cac-49cf-a55f-fbf215abebef {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.457435] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf81ea18-2b15-4327-94e1-ee773e900712 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.472811] env[63021]: DEBUG nova.compute.provider_tree [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 859.515406] env[63021]: DEBUG oslo_concurrency.lockutils [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Releasing lock "refresh_cache-6f4e6245-c0d9-46b4-baf5-5aca19e66da5" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.515896] env[63021]: DEBUG nova.compute.manager [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 859.516152] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 859.517117] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89392c14-fd7f-45e1-8812-bde8433251ed {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.526343] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 859.526601] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0ac825b7-6c46-4062-a6fe-72e01caa01e2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.535420] env[63021]: DEBUG oslo_vmware.api [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Waiting for the task: (returnval){ [ 859.535420] env[63021]: value = "task-1293785" [ 859.535420] env[63021]: _type = "Task" [ 859.535420] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.545643] env[63021]: DEBUG oslo_vmware.api [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293785, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.647290] env[63021]: DEBUG oslo_vmware.api [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]522dacaa-3b72-b2c8-725d-70a9f4b1ada4, 'name': SearchDatastore_Task, 'duration_secs': 0.011255} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.647669] env[63021]: DEBUG oslo_concurrency.lockutils [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.647912] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 859.648189] env[63021]: DEBUG oslo_concurrency.lockutils [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.648347] env[63021]: DEBUG oslo_concurrency.lockutils [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.648529] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 859.648824] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d0a218f2-453f-40ae-9465-9141c7cfb6f9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.660924] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 859.660924] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 859.661206] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28723113-5e0a-4d62-a63e-99bfd7229a1f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.672599] env[63021]: DEBUG oslo_vmware.api [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Waiting for the task: (returnval){ [ 859.672599] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52800890-020f-6770-76a6-a8949b594f4d" [ 859.672599] env[63021]: _type = "Task" [ 859.672599] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.682635] env[63021]: DEBUG oslo_vmware.api [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52800890-020f-6770-76a6-a8949b594f4d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.732951] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fee42473-b398-4575-8211-80c5332191e8 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "df856146-61ba-4de7-a5fd-d75263927163" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 79.025s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.836601] env[63021]: DEBUG nova.compute.manager [req-a6e5a491-9410-485c-a6f6-30fccbece295 req-547c1cb1-9824-4f60-83de-da58053a1311 service nova] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Received event network-vif-deleted-6d670553-cc52-4b24-8c5c-561d426a294a {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 859.836601] env[63021]: INFO nova.compute.manager [req-a6e5a491-9410-485c-a6f6-30fccbece295 req-547c1cb1-9824-4f60-83de-da58053a1311 service nova] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Neutron deleted interface 6d670553-cc52-4b24-8c5c-561d426a294a; detaching it from the instance and deleting it from the info cache [ 859.836601] env[63021]: DEBUG nova.network.neutron [req-a6e5a491-9410-485c-a6f6-30fccbece295 req-547c1cb1-9824-4f60-83de-da58053a1311 service nova] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.978073] env[63021]: DEBUG nova.scheduler.client.report [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 860.004638] env[63021]: DEBUG nova.network.neutron [-] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.048069] env[63021]: DEBUG oslo_vmware.api [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293785, 'name': PowerOffVM_Task, 'duration_secs': 0.15949} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.048069] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 860.048069] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 860.048948] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7e09c5e4-7c18-4acb-b750-1c4b7d292435 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.068226] env[63021]: DEBUG nova.compute.manager [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 860.082499] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 860.083914] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Deleting contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 860.083914] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Deleting the datastore file [datastore2] 6f4e6245-c0d9-46b4-baf5-5aca19e66da5 {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 860.083914] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f01b3eec-2f2e-4328-8cce-fd47965b069a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.096863] env[63021]: DEBUG oslo_vmware.api [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Waiting for the task: (returnval){ [ 860.096863] env[63021]: value = "task-1293787" [ 860.096863] env[63021]: _type = "Task" [ 860.096863] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.101139] env[63021]: DEBUG nova.virt.hardware [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 860.101476] env[63021]: DEBUG nova.virt.hardware [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 860.101660] env[63021]: DEBUG nova.virt.hardware [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 860.101896] env[63021]: DEBUG nova.virt.hardware [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 860.102112] env[63021]: DEBUG nova.virt.hardware [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 860.102281] env[63021]: DEBUG nova.virt.hardware [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 860.102552] env[63021]: DEBUG nova.virt.hardware [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 860.102757] env[63021]: DEBUG nova.virt.hardware [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 860.102949] env[63021]: DEBUG nova.virt.hardware [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 860.103183] env[63021]: DEBUG nova.virt.hardware [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 860.103394] env[63021]: DEBUG nova.virt.hardware [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 860.105065] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33cb4371-c3a2-4a85-8321-af87726c6568 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.122903] env[63021]: DEBUG oslo_vmware.api [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293787, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.124658] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6f03b6a-32b9-46e2-8196-4c8e2728a794 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.184121] env[63021]: DEBUG oslo_vmware.api [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52800890-020f-6770-76a6-a8949b594f4d, 'name': SearchDatastore_Task, 'duration_secs': 0.010949} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.184983] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1a5a3a8-e7d2-49cb-a05e-3ab980f29ceb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.192112] env[63021]: DEBUG oslo_vmware.api [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Waiting for the task: (returnval){ [ 860.192112] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52285b69-e934-ec63-ec7b-3a4e03967171" [ 860.192112] env[63021]: _type = "Task" [ 860.192112] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.203156] env[63021]: DEBUG oslo_vmware.api [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52285b69-e934-ec63-ec7b-3a4e03967171, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.235451] env[63021]: DEBUG nova.compute.manager [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 860.343539] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bd07c9b3-0b6f-497f-9ec0-82e1f5ec25c8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.354581] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f228482f-19ad-4db5-a4f4-e6f7801f927e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.390858] env[63021]: DEBUG nova.compute.manager [req-a6e5a491-9410-485c-a6f6-30fccbece295 req-547c1cb1-9824-4f60-83de-da58053a1311 service nova] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Detach interface failed, port_id=6d670553-cc52-4b24-8c5c-561d426a294a, reason: Instance 70f67905-7c60-433d-9ebc-d66fa44eb36e could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 860.486071] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.442s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.486628] env[63021]: DEBUG nova.compute.manager [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 860.491027] env[63021]: DEBUG oslo_concurrency.lockutils [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.231s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.491027] env[63021]: DEBUG nova.objects.instance [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Lazy-loading 'resources' on Instance uuid 9a6720ef-5307-44b7-b566-8d553c9b5384 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 860.507459] env[63021]: INFO nova.compute.manager [-] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Took 1.24 seconds to deallocate network for instance. [ 860.614883] env[63021]: DEBUG oslo_vmware.api [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Task: {'id': task-1293787, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.103012} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.615144] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 860.615393] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Deleted contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 860.615611] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 860.615808] env[63021]: INFO nova.compute.manager [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Took 1.10 seconds to destroy the instance on the hypervisor. [ 860.616110] env[63021]: DEBUG oslo.service.loopingcall [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 860.616350] env[63021]: DEBUG nova.compute.manager [-] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 860.616467] env[63021]: DEBUG nova.network.neutron [-] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 860.619535] env[63021]: DEBUG nova.network.neutron [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Successfully updated port: eff6f35a-0a13-4e0c-88b0-96b66b240f9b {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 860.624656] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bd0806d3-b3b3-409c-b331-faa1aa51b2bf tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "interface-36ec45ec-761b-4d62-a74f-e4d9a840ada0-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.624897] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bd0806d3-b3b3-409c-b331-faa1aa51b2bf tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "interface-36ec45ec-761b-4d62-a74f-e4d9a840ada0-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.625277] env[63021]: DEBUG nova.objects.instance [None req-bd0806d3-b3b3-409c-b331-faa1aa51b2bf tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lazy-loading 'flavor' on Instance uuid 36ec45ec-761b-4d62-a74f-e4d9a840ada0 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 860.633888] env[63021]: DEBUG nova.network.neutron [-] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 860.704095] env[63021]: DEBUG oslo_vmware.api [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52285b69-e934-ec63-ec7b-3a4e03967171, 'name': SearchDatastore_Task, 'duration_secs': 0.030075} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.704095] env[63021]: DEBUG oslo_concurrency.lockutils [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.704782] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 45c9e2be-eb19-4b83-b280-c9eeaddfccba/45c9e2be-eb19-4b83-b280-c9eeaddfccba.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 860.704782] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3c0f2ffd-4d75-4da6-a529-737ce3d19e99 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.713047] env[63021]: DEBUG oslo_vmware.api [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Waiting for the task: (returnval){ [ 860.713047] env[63021]: value = "task-1293788" [ 860.713047] env[63021]: _type = "Task" [ 860.713047] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.722695] env[63021]: DEBUG oslo_vmware.api [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1293788, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.762487] env[63021]: DEBUG oslo_concurrency.lockutils [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.996258] env[63021]: DEBUG nova.compute.utils [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 860.998027] env[63021]: DEBUG nova.compute.manager [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 860.998255] env[63021]: DEBUG nova.network.neutron [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 861.017916] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.051457] env[63021]: DEBUG nova.policy [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b427f7d1317341f38754a2b324f4f5aa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e238d5c1668540d2a9fffd3fd832d9b8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 861.122463] env[63021]: DEBUG oslo_concurrency.lockutils [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "refresh_cache-c328d2f7-3398-4f25-b11c-f464be7af8a1" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.122597] env[63021]: DEBUG oslo_concurrency.lockutils [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquired lock "refresh_cache-c328d2f7-3398-4f25-b11c-f464be7af8a1" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.122793] env[63021]: DEBUG nova.network.neutron [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 861.132817] env[63021]: DEBUG nova.objects.instance [None req-bd0806d3-b3b3-409c-b331-faa1aa51b2bf tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lazy-loading 'pci_requests' on Instance uuid 36ec45ec-761b-4d62-a74f-e4d9a840ada0 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 861.137953] env[63021]: DEBUG nova.network.neutron [-] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.229996] env[63021]: DEBUG oslo_vmware.api [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1293788, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.288212] env[63021]: DEBUG nova.compute.manager [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Stashing vm_state: active {{(pid=63021) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 861.386318] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5948f8e0-a495-4633-b9fd-9d460416a954 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.393377] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb6c06b3-b427-41b2-8237-4193c47b66bf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.437479] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a77e2ed-e2be-401c-86dc-999e3333a4aa {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.445818] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eda0005-9b53-43c1-9631-e38ed8ae3e58 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.462135] env[63021]: DEBUG nova.network.neutron [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Successfully created port: 7f87a718-8fb1-40d8-8273-ebd48f298224 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 861.464172] env[63021]: DEBUG nova.compute.provider_tree [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 861.501819] env[63021]: DEBUG nova.compute.manager [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 861.633322] env[63021]: DEBUG nova.objects.base [None req-bd0806d3-b3b3-409c-b331-faa1aa51b2bf tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Object Instance<36ec45ec-761b-4d62-a74f-e4d9a840ada0> lazy-loaded attributes: flavor,pci_requests {{(pid=63021) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 861.633586] env[63021]: DEBUG nova.network.neutron [None req-bd0806d3-b3b3-409c-b331-faa1aa51b2bf tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 861.640985] env[63021]: INFO nova.compute.manager [-] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Took 1.02 seconds to deallocate network for instance. [ 861.669521] env[63021]: DEBUG nova.network.neutron [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 861.726080] env[63021]: DEBUG oslo_vmware.api [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1293788, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.531497} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.726589] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 45c9e2be-eb19-4b83-b280-c9eeaddfccba/45c9e2be-eb19-4b83-b280-c9eeaddfccba.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 861.727133] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 861.727237] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-256db213-ca48-4166-a716-e9517c5115b4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.734387] env[63021]: DEBUG oslo_vmware.api [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Waiting for the task: (returnval){ [ 861.734387] env[63021]: value = "task-1293789" [ 861.734387] env[63021]: _type = "Task" [ 861.734387] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.748660] env[63021]: DEBUG oslo_vmware.api [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1293789, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.808652] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.824838] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bd0806d3-b3b3-409c-b331-faa1aa51b2bf tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "interface-36ec45ec-761b-4d62-a74f-e4d9a840ada0-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.200s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.868591] env[63021]: DEBUG nova.network.neutron [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Updating instance_info_cache with network_info: [{"id": "eff6f35a-0a13-4e0c-88b0-96b66b240f9b", "address": "fa:16:3e:27:3b:5e", "network": {"id": "9502e364-60e1-47ef-a1f7-f3e3ff45cf69", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-164717784-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5af87713ae6b477ca8a584aed287889d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeff6f35a-0a", "ovs_interfaceid": "eff6f35a-0a13-4e0c-88b0-96b66b240f9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.967377] env[63021]: DEBUG nova.scheduler.client.report [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 862.148211] env[63021]: DEBUG oslo_concurrency.lockutils [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.247324] env[63021]: DEBUG oslo_vmware.api [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1293789, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.263815} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.247619] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 862.248418] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7dbecee-051d-4ad0-af1d-c1bdf10e0bfd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.278231] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] 45c9e2be-eb19-4b83-b280-c9eeaddfccba/45c9e2be-eb19-4b83-b280-c9eeaddfccba.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 862.278231] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a8b2c072-1dfd-42a0-853e-68ae31fa52b6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.302055] env[63021]: DEBUG oslo_vmware.api [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Waiting for the task: (returnval){ [ 862.302055] env[63021]: value = "task-1293790" [ 862.302055] env[63021]: _type = "Task" [ 862.302055] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.311054] env[63021]: DEBUG oslo_vmware.api [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1293790, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.371550] env[63021]: DEBUG oslo_concurrency.lockutils [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Releasing lock "refresh_cache-c328d2f7-3398-4f25-b11c-f464be7af8a1" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.372112] env[63021]: DEBUG nova.compute.manager [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Instance network_info: |[{"id": "eff6f35a-0a13-4e0c-88b0-96b66b240f9b", "address": "fa:16:3e:27:3b:5e", "network": {"id": "9502e364-60e1-47ef-a1f7-f3e3ff45cf69", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-164717784-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5af87713ae6b477ca8a584aed287889d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeff6f35a-0a", "ovs_interfaceid": "eff6f35a-0a13-4e0c-88b0-96b66b240f9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 862.372447] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:27:3b:5e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eff6f35a-0a13-4e0c-88b0-96b66b240f9b', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 862.379848] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Creating folder: Project (5af87713ae6b477ca8a584aed287889d). Parent ref: group-v277447. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 862.380426] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3d5bd5d2-056e-4c18-a99e-17a87cdd447d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.393791] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Created folder: Project (5af87713ae6b477ca8a584aed287889d) in parent group-v277447. [ 862.393983] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Creating folder: Instances. Parent ref: group-v277516. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 862.394239] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-52ba07ba-796b-45d2-add1-ec66070cceee {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.404338] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Created folder: Instances in parent group-v277516. [ 862.404980] env[63021]: DEBUG oslo.service.loopingcall [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 862.404980] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 862.405961] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c5917cff-ea1a-4eb8-978e-61cf8e4817c4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.426276] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 862.426276] env[63021]: value = "task-1293793" [ 862.426276] env[63021]: _type = "Task" [ 862.426276] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.435799] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293793, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.474035] env[63021]: DEBUG oslo_concurrency.lockutils [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.984s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.477819] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.816s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.480035] env[63021]: INFO nova.compute.claims [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 862.502150] env[63021]: INFO nova.scheduler.client.report [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Deleted allocations for instance 9a6720ef-5307-44b7-b566-8d553c9b5384 [ 862.513299] env[63021]: DEBUG nova.compute.manager [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 862.547600] env[63021]: DEBUG nova.virt.hardware [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 862.547882] env[63021]: DEBUG nova.virt.hardware [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 862.548111] env[63021]: DEBUG nova.virt.hardware [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 862.548360] env[63021]: DEBUG nova.virt.hardware [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 862.548535] env[63021]: DEBUG nova.virt.hardware [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 862.548734] env[63021]: DEBUG nova.virt.hardware [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 862.549010] env[63021]: DEBUG nova.virt.hardware [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 862.549341] env[63021]: DEBUG nova.virt.hardware [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 862.549515] env[63021]: DEBUG nova.virt.hardware [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 862.549733] env[63021]: DEBUG nova.virt.hardware [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 862.549992] env[63021]: DEBUG nova.virt.hardware [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 862.551489] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3bb4b1a-b692-4a50-a44f-0fa273faa0f5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.561736] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbca6ebf-2f79-4b0a-befa-2a5c482456bc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.814869] env[63021]: DEBUG oslo_vmware.api [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1293790, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.939868] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293793, 'name': CreateVM_Task, 'duration_secs': 0.510397} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.940128] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 862.940828] env[63021]: DEBUG oslo_concurrency.lockutils [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.940994] env[63021]: DEBUG oslo_concurrency.lockutils [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.941381] env[63021]: DEBUG oslo_concurrency.lockutils [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 862.941653] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2967bfb4-7104-4067-97be-e136ad524591 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.947808] env[63021]: DEBUG oslo_vmware.api [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 862.947808] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]527f1a57-c1bc-daed-1e67-425d2474673b" [ 862.947808] env[63021]: _type = "Task" [ 862.947808] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.958314] env[63021]: DEBUG oslo_vmware.api [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]527f1a57-c1bc-daed-1e67-425d2474673b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.009958] env[63021]: DEBUG oslo_concurrency.lockutils [None req-303d7d1f-155a-480e-b2c0-f3ad88154661 tempest-ServersAdmin275Test-1922987333 tempest-ServersAdmin275Test-1922987333-project-member] Lock "9a6720ef-5307-44b7-b566-8d553c9b5384" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.458s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.053789] env[63021]: DEBUG nova.compute.manager [req-6a36ca82-b2e3-4f16-be22-1f31fc077b23 req-5a267507-7f95-4e94-9ea2-4b1e2b025910 service nova] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Received event network-vif-plugged-eff6f35a-0a13-4e0c-88b0-96b66b240f9b {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 863.054125] env[63021]: DEBUG oslo_concurrency.lockutils [req-6a36ca82-b2e3-4f16-be22-1f31fc077b23 req-5a267507-7f95-4e94-9ea2-4b1e2b025910 service nova] Acquiring lock "c328d2f7-3398-4f25-b11c-f464be7af8a1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.054394] env[63021]: DEBUG oslo_concurrency.lockutils [req-6a36ca82-b2e3-4f16-be22-1f31fc077b23 req-5a267507-7f95-4e94-9ea2-4b1e2b025910 service nova] Lock "c328d2f7-3398-4f25-b11c-f464be7af8a1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.054787] env[63021]: DEBUG oslo_concurrency.lockutils [req-6a36ca82-b2e3-4f16-be22-1f31fc077b23 req-5a267507-7f95-4e94-9ea2-4b1e2b025910 service nova] Lock "c328d2f7-3398-4f25-b11c-f464be7af8a1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.055108] env[63021]: DEBUG nova.compute.manager [req-6a36ca82-b2e3-4f16-be22-1f31fc077b23 req-5a267507-7f95-4e94-9ea2-4b1e2b025910 service nova] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] No waiting events found dispatching network-vif-plugged-eff6f35a-0a13-4e0c-88b0-96b66b240f9b {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 863.055664] env[63021]: WARNING nova.compute.manager [req-6a36ca82-b2e3-4f16-be22-1f31fc077b23 req-5a267507-7f95-4e94-9ea2-4b1e2b025910 service nova] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Received unexpected event network-vif-plugged-eff6f35a-0a13-4e0c-88b0-96b66b240f9b for instance with vm_state building and task_state spawning. [ 863.055875] env[63021]: DEBUG nova.compute.manager [req-6a36ca82-b2e3-4f16-be22-1f31fc077b23 req-5a267507-7f95-4e94-9ea2-4b1e2b025910 service nova] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Received event network-changed-eff6f35a-0a13-4e0c-88b0-96b66b240f9b {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 863.056050] env[63021]: DEBUG nova.compute.manager [req-6a36ca82-b2e3-4f16-be22-1f31fc077b23 req-5a267507-7f95-4e94-9ea2-4b1e2b025910 service nova] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Refreshing instance network info cache due to event network-changed-eff6f35a-0a13-4e0c-88b0-96b66b240f9b. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 863.056244] env[63021]: DEBUG oslo_concurrency.lockutils [req-6a36ca82-b2e3-4f16-be22-1f31fc077b23 req-5a267507-7f95-4e94-9ea2-4b1e2b025910 service nova] Acquiring lock "refresh_cache-c328d2f7-3398-4f25-b11c-f464be7af8a1" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.056380] env[63021]: DEBUG oslo_concurrency.lockutils [req-6a36ca82-b2e3-4f16-be22-1f31fc077b23 req-5a267507-7f95-4e94-9ea2-4b1e2b025910 service nova] Acquired lock "refresh_cache-c328d2f7-3398-4f25-b11c-f464be7af8a1" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.056535] env[63021]: DEBUG nova.network.neutron [req-6a36ca82-b2e3-4f16-be22-1f31fc077b23 req-5a267507-7f95-4e94-9ea2-4b1e2b025910 service nova] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Refreshing network info cache for port eff6f35a-0a13-4e0c-88b0-96b66b240f9b {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 863.192208] env[63021]: DEBUG nova.compute.manager [req-5eb7164d-4276-4e94-9073-3d011858295c req-b3223587-155d-4f89-bb87-fc7d9a67ea4b service nova] [instance: df856146-61ba-4de7-a5fd-d75263927163] Received event network-changed-a43dcfd2-397e-44fe-bd1c-b6116c25a28a {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 863.192908] env[63021]: DEBUG nova.compute.manager [req-5eb7164d-4276-4e94-9073-3d011858295c req-b3223587-155d-4f89-bb87-fc7d9a67ea4b service nova] [instance: df856146-61ba-4de7-a5fd-d75263927163] Refreshing instance network info cache due to event network-changed-a43dcfd2-397e-44fe-bd1c-b6116c25a28a. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 863.193838] env[63021]: DEBUG oslo_concurrency.lockutils [req-5eb7164d-4276-4e94-9073-3d011858295c req-b3223587-155d-4f89-bb87-fc7d9a67ea4b service nova] Acquiring lock "refresh_cache-df856146-61ba-4de7-a5fd-d75263927163" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.197021] env[63021]: DEBUG oslo_concurrency.lockutils [req-5eb7164d-4276-4e94-9073-3d011858295c req-b3223587-155d-4f89-bb87-fc7d9a67ea4b service nova] Acquired lock "refresh_cache-df856146-61ba-4de7-a5fd-d75263927163" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.197021] env[63021]: DEBUG nova.network.neutron [req-5eb7164d-4276-4e94-9073-3d011858295c req-b3223587-155d-4f89-bb87-fc7d9a67ea4b service nova] [instance: df856146-61ba-4de7-a5fd-d75263927163] Refreshing network info cache for port a43dcfd2-397e-44fe-bd1c-b6116c25a28a {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 863.244084] env[63021]: DEBUG nova.network.neutron [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Successfully updated port: 7f87a718-8fb1-40d8-8273-ebd48f298224 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 863.315024] env[63021]: DEBUG oslo_vmware.api [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1293790, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.461804] env[63021]: DEBUG oslo_vmware.api [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]527f1a57-c1bc-daed-1e67-425d2474673b, 'name': SearchDatastore_Task, 'duration_secs': 0.010183} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.462602] env[63021]: DEBUG oslo_concurrency.lockutils [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.462965] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 863.463321] env[63021]: DEBUG oslo_concurrency.lockutils [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.463580] env[63021]: DEBUG oslo_concurrency.lockutils [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.463867] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 863.464417] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2072731f-740f-4f8b-9e64-6e5b5687eabf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.476785] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 863.476785] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 863.476785] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd2e4f29-af2c-4c32-a7af-a515b330df55 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.482174] env[63021]: DEBUG oslo_vmware.api [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 863.482174] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5250198e-634b-ca76-cdaa-bcfd81e251e9" [ 863.482174] env[63021]: _type = "Task" [ 863.482174] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.493847] env[63021]: DEBUG oslo_vmware.api [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5250198e-634b-ca76-cdaa-bcfd81e251e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.747202] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "refresh_cache-16b21798-17e8-4aeb-affa-57eae31d5dd3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.747566] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquired lock "refresh_cache-16b21798-17e8-4aeb-affa-57eae31d5dd3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.747846] env[63021]: DEBUG nova.network.neutron [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 863.818241] env[63021]: DEBUG oslo_vmware.api [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1293790, 'name': ReconfigVM_Task, 'duration_secs': 1.445626} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.818563] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Reconfigured VM instance instance-00000044 to attach disk [datastore1] 45c9e2be-eb19-4b83-b280-c9eeaddfccba/45c9e2be-eb19-4b83-b280-c9eeaddfccba.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 863.819240] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0be4c37b-9476-4512-bc5d-ee6150f05cef {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.829168] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d97f2a0e-17a8-4278-bf37-9412dc50bb66 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.838659] env[63021]: DEBUG oslo_vmware.api [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Waiting for the task: (returnval){ [ 863.838659] env[63021]: value = "task-1293794" [ 863.838659] env[63021]: _type = "Task" [ 863.838659] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.849254] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c16ff284-c487-4758-aa1e-cb58b1da8580 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.852735] env[63021]: DEBUG oslo_vmware.api [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1293794, 'name': Rename_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.889039] env[63021]: DEBUG nova.network.neutron [req-6a36ca82-b2e3-4f16-be22-1f31fc077b23 req-5a267507-7f95-4e94-9ea2-4b1e2b025910 service nova] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Updated VIF entry in instance network info cache for port eff6f35a-0a13-4e0c-88b0-96b66b240f9b. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 863.889039] env[63021]: DEBUG nova.network.neutron [req-6a36ca82-b2e3-4f16-be22-1f31fc077b23 req-5a267507-7f95-4e94-9ea2-4b1e2b025910 service nova] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Updating instance_info_cache with network_info: [{"id": "eff6f35a-0a13-4e0c-88b0-96b66b240f9b", "address": "fa:16:3e:27:3b:5e", "network": {"id": "9502e364-60e1-47ef-a1f7-f3e3ff45cf69", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-164717784-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5af87713ae6b477ca8a584aed287889d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeff6f35a-0a", "ovs_interfaceid": "eff6f35a-0a13-4e0c-88b0-96b66b240f9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.894736] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1b1fae9-e9f5-49b0-bb3a-8fd27a1650ba {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.904448] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a68cf9b-6f3d-4e5a-99c8-02d68cc8018c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.921528] env[63021]: DEBUG nova.compute.provider_tree [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 863.995745] env[63021]: DEBUG oslo_vmware.api [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5250198e-634b-ca76-cdaa-bcfd81e251e9, 'name': SearchDatastore_Task, 'duration_secs': 0.014155} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.999049] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-97e01f07-a848-4448-961e-e10e920bdd35 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.005585] env[63021]: DEBUG oslo_vmware.api [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 864.005585] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5274844b-ffc5-9168-a4ba-1a0381e42b2d" [ 864.005585] env[63021]: _type = "Task" [ 864.005585] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.014450] env[63021]: DEBUG oslo_vmware.api [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5274844b-ffc5-9168-a4ba-1a0381e42b2d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.042622] env[63021]: DEBUG nova.network.neutron [req-5eb7164d-4276-4e94-9073-3d011858295c req-b3223587-155d-4f89-bb87-fc7d9a67ea4b service nova] [instance: df856146-61ba-4de7-a5fd-d75263927163] Updated VIF entry in instance network info cache for port a43dcfd2-397e-44fe-bd1c-b6116c25a28a. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 864.042974] env[63021]: DEBUG nova.network.neutron [req-5eb7164d-4276-4e94-9073-3d011858295c req-b3223587-155d-4f89-bb87-fc7d9a67ea4b service nova] [instance: df856146-61ba-4de7-a5fd-d75263927163] Updating instance_info_cache with network_info: [{"id": "a43dcfd2-397e-44fe-bd1c-b6116c25a28a", "address": "fa:16:3e:c8:73:42", "network": {"id": "7258df3e-b1eb-4da6-ab52-5dadd2f87b51", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1084458843-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.144", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5661c44b534b4a7e9452e914fc80d66e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4f04accd-4a9f-41a1-b327-c96095e1ade4", "external-id": "nsx-vlan-transportzone-947", "segmentation_id": 947, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa43dcfd2-39", "ovs_interfaceid": "a43dcfd2-397e-44fe-bd1c-b6116c25a28a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.057569] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "interface-36ec45ec-761b-4d62-a74f-e4d9a840ada0-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.057816] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "interface-36ec45ec-761b-4d62-a74f-e4d9a840ada0-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.058162] env[63021]: DEBUG nova.objects.instance [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lazy-loading 'flavor' on Instance uuid 36ec45ec-761b-4d62-a74f-e4d9a840ada0 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 864.287842] env[63021]: DEBUG nova.network.neutron [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 864.353964] env[63021]: DEBUG oslo_vmware.api [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1293794, 'name': Rename_Task, 'duration_secs': 0.182047} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.353964] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 864.354410] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-75105b51-b9ff-48b2-91b3-6b691db3171d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.362589] env[63021]: DEBUG oslo_vmware.api [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Waiting for the task: (returnval){ [ 864.362589] env[63021]: value = "task-1293795" [ 864.362589] env[63021]: _type = "Task" [ 864.362589] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.374170] env[63021]: DEBUG oslo_vmware.api [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1293795, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.399114] env[63021]: DEBUG oslo_concurrency.lockutils [req-6a36ca82-b2e3-4f16-be22-1f31fc077b23 req-5a267507-7f95-4e94-9ea2-4b1e2b025910 service nova] Releasing lock "refresh_cache-c328d2f7-3398-4f25-b11c-f464be7af8a1" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.426916] env[63021]: DEBUG nova.scheduler.client.report [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 864.517274] env[63021]: DEBUG oslo_vmware.api [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5274844b-ffc5-9168-a4ba-1a0381e42b2d, 'name': SearchDatastore_Task, 'duration_secs': 0.011141} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.517551] env[63021]: DEBUG oslo_concurrency.lockutils [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.517788] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] c328d2f7-3398-4f25-b11c-f464be7af8a1/c328d2f7-3398-4f25-b11c-f464be7af8a1.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 864.518060] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ca7a8877-c08f-4c50-a1a3-29311328a1dc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.526234] env[63021]: DEBUG oslo_vmware.api [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 864.526234] env[63021]: value = "task-1293796" [ 864.526234] env[63021]: _type = "Task" [ 864.526234] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.534993] env[63021]: DEBUG oslo_vmware.api [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1293796, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.545981] env[63021]: DEBUG oslo_concurrency.lockutils [req-5eb7164d-4276-4e94-9073-3d011858295c req-b3223587-155d-4f89-bb87-fc7d9a67ea4b service nova] Releasing lock "refresh_cache-df856146-61ba-4de7-a5fd-d75263927163" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.669174] env[63021]: DEBUG nova.network.neutron [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Updating instance_info_cache with network_info: [{"id": "7f87a718-8fb1-40d8-8273-ebd48f298224", "address": "fa:16:3e:45:fa:52", "network": {"id": "81917a55-4e41-4898-9f4e-4fced52b5a86", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1793837736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e238d5c1668540d2a9fffd3fd832d9b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f87a718-8f", "ovs_interfaceid": "7f87a718-8fb1-40d8-8273-ebd48f298224", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.677436] env[63021]: DEBUG nova.objects.instance [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lazy-loading 'pci_requests' on Instance uuid 36ec45ec-761b-4d62-a74f-e4d9a840ada0 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 864.875557] env[63021]: DEBUG oslo_vmware.api [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1293795, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.932030] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.455s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.932855] env[63021]: DEBUG nova.compute.manager [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 864.936303] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.179s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.937042] env[63021]: DEBUG nova.objects.instance [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lazy-loading 'resources' on Instance uuid 2209d709-7a5b-42e2-be93-b4fd436b6290 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 865.044304] env[63021]: DEBUG oslo_vmware.api [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1293796, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.172667] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Releasing lock "refresh_cache-16b21798-17e8-4aeb-affa-57eae31d5dd3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.173067] env[63021]: DEBUG nova.compute.manager [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Instance network_info: |[{"id": "7f87a718-8fb1-40d8-8273-ebd48f298224", "address": "fa:16:3e:45:fa:52", "network": {"id": "81917a55-4e41-4898-9f4e-4fced52b5a86", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1793837736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e238d5c1668540d2a9fffd3fd832d9b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f87a718-8f", "ovs_interfaceid": "7f87a718-8fb1-40d8-8273-ebd48f298224", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 865.173532] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:45:fa:52', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '535b175f-71d3-4226-81ab-ca253f27fedd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7f87a718-8fb1-40d8-8273-ebd48f298224', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 865.181323] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Creating folder: Project (e238d5c1668540d2a9fffd3fd832d9b8). Parent ref: group-v277447. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 865.181992] env[63021]: DEBUG nova.objects.base [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Object Instance<36ec45ec-761b-4d62-a74f-e4d9a840ada0> lazy-loaded attributes: flavor,pci_requests {{(pid=63021) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 865.182221] env[63021]: DEBUG nova.network.neutron [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 865.184412] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-df613de2-0af0-4d2c-969a-64cc67fcb84b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.201221] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Created folder: Project (e238d5c1668540d2a9fffd3fd832d9b8) in parent group-v277447. [ 865.202256] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Creating folder: Instances. Parent ref: group-v277519. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 865.202256] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9594a83e-f8d2-4a35-ada5-5318f320b540 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.219059] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Created folder: Instances in parent group-v277519. [ 865.219243] env[63021]: DEBUG oslo.service.loopingcall [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 865.219561] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 865.219862] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-171cb8ca-59e6-41ad-90db-1a34a4382d6e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.243040] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 865.243040] env[63021]: value = "task-1293799" [ 865.243040] env[63021]: _type = "Task" [ 865.243040] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.253973] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293799, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.256308] env[63021]: DEBUG nova.policy [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6ea19453dbdb4251a93dde2576aeff64', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '58693d1e991249f28914a39761cc2627', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 865.378392] env[63021]: DEBUG oslo_vmware.api [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1293795, 'name': PowerOnVM_Task, 'duration_secs': 0.563534} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.378739] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 865.378951] env[63021]: INFO nova.compute.manager [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Took 10.78 seconds to spawn the instance on the hypervisor. [ 865.379148] env[63021]: DEBUG nova.compute.manager [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 865.384219] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9204c0db-987d-48f4-ac50-0df05e9bbc06 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.451140] env[63021]: DEBUG nova.compute.utils [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 865.454119] env[63021]: DEBUG nova.compute.manager [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 865.456672] env[63021]: DEBUG nova.network.neutron [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 865.541273] env[63021]: DEBUG oslo_vmware.api [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1293796, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.702788} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.544261] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] c328d2f7-3398-4f25-b11c-f464be7af8a1/c328d2f7-3398-4f25-b11c-f464be7af8a1.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 865.544706] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 865.545331] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-59721f84-1a85-4d84-92f1-4fb763c624e9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.553039] env[63021]: DEBUG nova.policy [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1e133aeda3249ec98dbdc53f869d719', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7331caed42034dfc81c79cfabbe0829e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 865.557636] env[63021]: DEBUG oslo_vmware.api [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 865.557636] env[63021]: value = "task-1293800" [ 865.557636] env[63021]: _type = "Task" [ 865.557636] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.574423] env[63021]: DEBUG oslo_vmware.api [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1293800, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.716933] env[63021]: DEBUG nova.network.neutron [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Successfully created port: d92edf1f-d15d-4f56-9da3-92cbe44b6709 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 865.757877] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293799, 'name': CreateVM_Task} progress is 25%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.839684] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b2ba773-c110-428a-8f26-e5b82836d236 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.850838] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c43eaba4-eb22-418f-9729-d42a93d82ccb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.894300] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f6924bf-a087-4fc9-a9ef-1a91c840a96d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.905720] env[63021]: INFO nova.compute.manager [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Took 40.11 seconds to build instance. [ 865.911141] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-466a0b07-c394-4743-b054-028ef4e8b117 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.930537] env[63021]: DEBUG nova.compute.provider_tree [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 865.962472] env[63021]: DEBUG nova.compute.manager [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 866.073075] env[63021]: DEBUG oslo_vmware.api [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1293800, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.092093} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.073481] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 866.074302] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b7fdaf0-6dc5-4ce7-8ac6-a868ebbbed12 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.099231] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] c328d2f7-3398-4f25-b11c-f464be7af8a1/c328d2f7-3398-4f25-b11c-f464be7af8a1.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 866.100283] env[63021]: DEBUG nova.network.neutron [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Successfully created port: 3dc162c2-36e3-4bb0-b7a4-3116c11c145a {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 866.102346] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-50a27065-9537-41f2-8b5f-31e917a63791 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.124620] env[63021]: DEBUG oslo_vmware.api [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 866.124620] env[63021]: value = "task-1293801" [ 866.124620] env[63021]: _type = "Task" [ 866.124620] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.134054] env[63021]: DEBUG oslo_vmware.api [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1293801, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.255780] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293799, 'name': CreateVM_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.411353] env[63021]: DEBUG oslo_concurrency.lockutils [None req-df969cb3-4d54-494d-b14f-eaf331adeb66 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lock "45c9e2be-eb19-4b83-b280-c9eeaddfccba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.242s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.434249] env[63021]: DEBUG nova.scheduler.client.report [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 866.636672] env[63021]: DEBUG oslo_vmware.api [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1293801, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.757228] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293799, 'name': CreateVM_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.914617] env[63021]: DEBUG nova.compute.manager [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 866.939691] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.003s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.944168] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.129s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.944594] env[63021]: DEBUG nova.objects.instance [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lazy-loading 'resources' on Instance uuid 36d26354-6127-4557-acb0-8743e27ff1ff {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 866.966154] env[63021]: INFO nova.scheduler.client.report [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Deleted allocations for instance 2209d709-7a5b-42e2-be93-b4fd436b6290 [ 866.973093] env[63021]: DEBUG nova.compute.manager [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 867.009931] env[63021]: DEBUG nova.virt.hardware [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 867.010201] env[63021]: DEBUG nova.virt.hardware [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 867.010353] env[63021]: DEBUG nova.virt.hardware [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 867.011036] env[63021]: DEBUG nova.virt.hardware [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 867.011036] env[63021]: DEBUG nova.virt.hardware [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 867.011036] env[63021]: DEBUG nova.virt.hardware [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 867.011036] env[63021]: DEBUG nova.virt.hardware [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 867.011321] env[63021]: DEBUG nova.virt.hardware [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 867.011321] env[63021]: DEBUG nova.virt.hardware [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 867.011671] env[63021]: DEBUG nova.virt.hardware [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 867.011671] env[63021]: DEBUG nova.virt.hardware [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 867.012522] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48eec527-2f52-4ae4-ae10-de0fd2288954 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.021572] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-795524ec-3caf-4ad8-88db-1212cfc28cd3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.135805] env[63021]: DEBUG oslo_vmware.api [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1293801, 'name': ReconfigVM_Task, 'duration_secs': 0.537681} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.136099] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Reconfigured VM instance instance-00000045 to attach disk [datastore1] c328d2f7-3398-4f25-b11c-f464be7af8a1/c328d2f7-3398-4f25-b11c-f464be7af8a1.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 867.136745] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6306e8c7-7a81-4668-b91d-b2bdfca4ad8f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.144312] env[63021]: DEBUG oslo_vmware.api [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 867.144312] env[63021]: value = "task-1293802" [ 867.144312] env[63021]: _type = "Task" [ 867.144312] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.153015] env[63021]: DEBUG oslo_vmware.api [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1293802, 'name': Rename_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.256693] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293799, 'name': CreateVM_Task, 'duration_secs': 1.764336} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.256899] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 867.257618] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.257784] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.258107] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 867.258363] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5d2d3ab-9ede-4f0e-8df8-3d8272eb3be4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.264251] env[63021]: DEBUG oslo_vmware.api [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 867.264251] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52a24776-d3a5-05a2-da1c-4b35972116a5" [ 867.264251] env[63021]: _type = "Task" [ 867.264251] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.273613] env[63021]: DEBUG oslo_vmware.api [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52a24776-d3a5-05a2-da1c-4b35972116a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.313855] env[63021]: DEBUG nova.compute.manager [req-54f099fb-6a8a-4796-9775-7dbf1edef9e3 req-1ba9c263-6de0-4475-b642-cbaa22f0468d service nova] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Received event network-vif-plugged-7f87a718-8fb1-40d8-8273-ebd48f298224 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 867.313855] env[63021]: DEBUG oslo_concurrency.lockutils [req-54f099fb-6a8a-4796-9775-7dbf1edef9e3 req-1ba9c263-6de0-4475-b642-cbaa22f0468d service nova] Acquiring lock "16b21798-17e8-4aeb-affa-57eae31d5dd3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.313855] env[63021]: DEBUG oslo_concurrency.lockutils [req-54f099fb-6a8a-4796-9775-7dbf1edef9e3 req-1ba9c263-6de0-4475-b642-cbaa22f0468d service nova] Lock "16b21798-17e8-4aeb-affa-57eae31d5dd3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.313855] env[63021]: DEBUG oslo_concurrency.lockutils [req-54f099fb-6a8a-4796-9775-7dbf1edef9e3 req-1ba9c263-6de0-4475-b642-cbaa22f0468d service nova] Lock "16b21798-17e8-4aeb-affa-57eae31d5dd3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.313855] env[63021]: DEBUG nova.compute.manager [req-54f099fb-6a8a-4796-9775-7dbf1edef9e3 req-1ba9c263-6de0-4475-b642-cbaa22f0468d service nova] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] No waiting events found dispatching network-vif-plugged-7f87a718-8fb1-40d8-8273-ebd48f298224 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 867.314452] env[63021]: WARNING nova.compute.manager [req-54f099fb-6a8a-4796-9775-7dbf1edef9e3 req-1ba9c263-6de0-4475-b642-cbaa22f0468d service nova] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Received unexpected event network-vif-plugged-7f87a718-8fb1-40d8-8273-ebd48f298224 for instance with vm_state building and task_state spawning. [ 867.314452] env[63021]: DEBUG nova.compute.manager [req-54f099fb-6a8a-4796-9775-7dbf1edef9e3 req-1ba9c263-6de0-4475-b642-cbaa22f0468d service nova] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Received event network-changed-7f87a718-8fb1-40d8-8273-ebd48f298224 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 867.314452] env[63021]: DEBUG nova.compute.manager [req-54f099fb-6a8a-4796-9775-7dbf1edef9e3 req-1ba9c263-6de0-4475-b642-cbaa22f0468d service nova] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Refreshing instance network info cache due to event network-changed-7f87a718-8fb1-40d8-8273-ebd48f298224. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 867.314580] env[63021]: DEBUG oslo_concurrency.lockutils [req-54f099fb-6a8a-4796-9775-7dbf1edef9e3 req-1ba9c263-6de0-4475-b642-cbaa22f0468d service nova] Acquiring lock "refresh_cache-16b21798-17e8-4aeb-affa-57eae31d5dd3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.314735] env[63021]: DEBUG oslo_concurrency.lockutils [req-54f099fb-6a8a-4796-9775-7dbf1edef9e3 req-1ba9c263-6de0-4475-b642-cbaa22f0468d service nova] Acquired lock "refresh_cache-16b21798-17e8-4aeb-affa-57eae31d5dd3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.314956] env[63021]: DEBUG nova.network.neutron [req-54f099fb-6a8a-4796-9775-7dbf1edef9e3 req-1ba9c263-6de0-4475-b642-cbaa22f0468d service nova] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Refreshing network info cache for port 7f87a718-8fb1-40d8-8273-ebd48f298224 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 867.435895] env[63021]: DEBUG nova.network.neutron [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Successfully updated port: d92edf1f-d15d-4f56-9da3-92cbe44b6709 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 867.444329] env[63021]: DEBUG oslo_concurrency.lockutils [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.473849] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cae6dd9b-162c-4a3d-9cb7-8deb2b506325 tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lock "2209d709-7a5b-42e2-be93-b4fd436b6290" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.630s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.492773] env[63021]: DEBUG nova.compute.manager [req-4c418585-305b-45d0-984e-0f68035680fc req-b49f813e-035f-4674-8c7a-83ab6f3addb8 service nova] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Received event network-changed-e581fd41-04b7-4081-a7be-c6389f4d623f {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 867.492978] env[63021]: DEBUG nova.compute.manager [req-4c418585-305b-45d0-984e-0f68035680fc req-b49f813e-035f-4674-8c7a-83ab6f3addb8 service nova] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Refreshing instance network info cache due to event network-changed-e581fd41-04b7-4081-a7be-c6389f4d623f. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 867.493285] env[63021]: DEBUG oslo_concurrency.lockutils [req-4c418585-305b-45d0-984e-0f68035680fc req-b49f813e-035f-4674-8c7a-83ab6f3addb8 service nova] Acquiring lock "refresh_cache-7787f6bb-e855-476f-9146-2a648932429d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.493510] env[63021]: DEBUG oslo_concurrency.lockutils [req-4c418585-305b-45d0-984e-0f68035680fc req-b49f813e-035f-4674-8c7a-83ab6f3addb8 service nova] Acquired lock "refresh_cache-7787f6bb-e855-476f-9146-2a648932429d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.493643] env[63021]: DEBUG nova.network.neutron [req-4c418585-305b-45d0-984e-0f68035680fc req-b49f813e-035f-4674-8c7a-83ab6f3addb8 service nova] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Refreshing network info cache for port e581fd41-04b7-4081-a7be-c6389f4d623f {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 867.658270] env[63021]: DEBUG oslo_vmware.api [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1293802, 'name': Rename_Task, 'duration_secs': 0.199982} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.658579] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 867.658858] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2abf094d-f391-41ff-bca1-bb1b4976fb03 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.666797] env[63021]: DEBUG oslo_vmware.api [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 867.666797] env[63021]: value = "task-1293803" [ 867.666797] env[63021]: _type = "Task" [ 867.666797] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.675720] env[63021]: DEBUG oslo_vmware.api [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1293803, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.692059] env[63021]: DEBUG nova.network.neutron [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Successfully updated port: 3dc162c2-36e3-4bb0-b7a4-3116c11c145a {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 867.773490] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08cea191-343e-43cb-a94a-26a3a6bab19b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.780169] env[63021]: DEBUG oslo_vmware.api [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52a24776-d3a5-05a2-da1c-4b35972116a5, 'name': SearchDatastore_Task, 'duration_secs': 0.011893} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.780887] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.781209] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 867.781409] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.781564] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.781750] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 867.782033] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0713e2c9-8aed-40c9-b4f6-354145fa14c7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.787744] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfaa26a7-2aa3-4473-9168-4738cb131055 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.794140] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 867.794392] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 867.820190] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6888e6a0-fd76-48ad-b83b-a24caaf89501 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.825746] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d967b55-a7e4-4d73-95d5-f2b82fb23319 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.837278] env[63021]: DEBUG oslo_vmware.api [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 867.837278] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52931b05-783a-204b-e91b-17a284239a85" [ 867.837278] env[63021]: _type = "Task" [ 867.837278] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.840741] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e4a6481-d6e4-47a8-8dfb-2182f9233440 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.852571] env[63021]: DEBUG oslo_vmware.api [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52931b05-783a-204b-e91b-17a284239a85, 'name': SearchDatastore_Task, 'duration_secs': 0.012375} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.862747] env[63021]: DEBUG nova.compute.provider_tree [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 867.867021] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9066e9d-cb64-4b22-bdc0-7d4e43f546b2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.873610] env[63021]: DEBUG oslo_vmware.api [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 867.873610] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52219a97-0eba-4cb2-9b53-a8ac0a25b9d3" [ 867.873610] env[63021]: _type = "Task" [ 867.873610] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.884196] env[63021]: DEBUG oslo_vmware.api [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52219a97-0eba-4cb2-9b53-a8ac0a25b9d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.940062] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "refresh_cache-36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.940392] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquired lock "refresh_cache-36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.940510] env[63021]: DEBUG nova.network.neutron [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 868.102020] env[63021]: DEBUG nova.network.neutron [req-54f099fb-6a8a-4796-9775-7dbf1edef9e3 req-1ba9c263-6de0-4475-b642-cbaa22f0468d service nova] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Updated VIF entry in instance network info cache for port 7f87a718-8fb1-40d8-8273-ebd48f298224. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 868.102423] env[63021]: DEBUG nova.network.neutron [req-54f099fb-6a8a-4796-9775-7dbf1edef9e3 req-1ba9c263-6de0-4475-b642-cbaa22f0468d service nova] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Updating instance_info_cache with network_info: [{"id": "7f87a718-8fb1-40d8-8273-ebd48f298224", "address": "fa:16:3e:45:fa:52", "network": {"id": "81917a55-4e41-4898-9f4e-4fced52b5a86", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1793837736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e238d5c1668540d2a9fffd3fd832d9b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f87a718-8f", "ovs_interfaceid": "7f87a718-8fb1-40d8-8273-ebd48f298224", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.177809] env[63021]: DEBUG oslo_vmware.api [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1293803, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.198460] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "refresh_cache-881c2fb8-5f8c-48c2-a173-8fab949f24f7" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.198460] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquired lock "refresh_cache-881c2fb8-5f8c-48c2-a173-8fab949f24f7" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.198460] env[63021]: DEBUG nova.network.neutron [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 868.265505] env[63021]: DEBUG nova.network.neutron [req-4c418585-305b-45d0-984e-0f68035680fc req-b49f813e-035f-4674-8c7a-83ab6f3addb8 service nova] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Updated VIF entry in instance network info cache for port e581fd41-04b7-4081-a7be-c6389f4d623f. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 868.265505] env[63021]: DEBUG nova.network.neutron [req-4c418585-305b-45d0-984e-0f68035680fc req-b49f813e-035f-4674-8c7a-83ab6f3addb8 service nova] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Updating instance_info_cache with network_info: [{"id": "e581fd41-04b7-4081-a7be-c6389f4d623f", "address": "fa:16:3e:5b:84:48", "network": {"id": "6d99496e-4e9f-4e5d-81a4-305f91e927cc", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-851847769-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.199", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "70e912b7ac9144c793323eaf26287ba1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f969bd9-e040-4b9b-85b2-7c61231584ad", "external-id": "nsx-vlan-transportzone-995", "segmentation_id": 995, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape581fd41-04", "ovs_interfaceid": "e581fd41-04b7-4081-a7be-c6389f4d623f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.370143] env[63021]: DEBUG nova.scheduler.client.report [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 868.386732] env[63021]: DEBUG oslo_vmware.api [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52219a97-0eba-4cb2-9b53-a8ac0a25b9d3, 'name': SearchDatastore_Task, 'duration_secs': 0.025875} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.387923] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.388130] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 16b21798-17e8-4aeb-affa-57eae31d5dd3/16b21798-17e8-4aeb-affa-57eae31d5dd3.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 868.388416] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ce22ab07-5a97-4812-a32e-7427bdf39fc7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.396175] env[63021]: DEBUG oslo_vmware.api [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 868.396175] env[63021]: value = "task-1293804" [ 868.396175] env[63021]: _type = "Task" [ 868.396175] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.405828] env[63021]: DEBUG oslo_vmware.api [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293804, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.479101] env[63021]: WARNING nova.network.neutron [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] 2c4f9111-2f39-4bd0-95bc-09b96d535ea3 already exists in list: networks containing: ['2c4f9111-2f39-4bd0-95bc-09b96d535ea3']. ignoring it [ 868.605514] env[63021]: DEBUG oslo_concurrency.lockutils [req-54f099fb-6a8a-4796-9775-7dbf1edef9e3 req-1ba9c263-6de0-4475-b642-cbaa22f0468d service nova] Releasing lock "refresh_cache-16b21798-17e8-4aeb-affa-57eae31d5dd3" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.677804] env[63021]: DEBUG oslo_vmware.api [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1293803, 'name': PowerOnVM_Task, 'duration_secs': 0.724416} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.679908] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 868.680931] env[63021]: INFO nova.compute.manager [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Took 8.61 seconds to spawn the instance on the hypervisor. [ 868.680931] env[63021]: DEBUG nova.compute.manager [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 868.681074] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbe6868d-cb54-4080-90bf-ac6350c95afd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.741297] env[63021]: DEBUG nova.network.neutron [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 868.768466] env[63021]: DEBUG oslo_concurrency.lockutils [req-4c418585-305b-45d0-984e-0f68035680fc req-b49f813e-035f-4674-8c7a-83ab6f3addb8 service nova] Releasing lock "refresh_cache-7787f6bb-e855-476f-9146-2a648932429d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.768747] env[63021]: DEBUG nova.compute.manager [req-4c418585-305b-45d0-984e-0f68035680fc req-b49f813e-035f-4674-8c7a-83ab6f3addb8 service nova] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Received event network-changed-e581fd41-04b7-4081-a7be-c6389f4d623f {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 868.768917] env[63021]: DEBUG nova.compute.manager [req-4c418585-305b-45d0-984e-0f68035680fc req-b49f813e-035f-4674-8c7a-83ab6f3addb8 service nova] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Refreshing instance network info cache due to event network-changed-e581fd41-04b7-4081-a7be-c6389f4d623f. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 868.769129] env[63021]: DEBUG oslo_concurrency.lockutils [req-4c418585-305b-45d0-984e-0f68035680fc req-b49f813e-035f-4674-8c7a-83ab6f3addb8 service nova] Acquiring lock "refresh_cache-7787f6bb-e855-476f-9146-2a648932429d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.769273] env[63021]: DEBUG oslo_concurrency.lockutils [req-4c418585-305b-45d0-984e-0f68035680fc req-b49f813e-035f-4674-8c7a-83ab6f3addb8 service nova] Acquired lock "refresh_cache-7787f6bb-e855-476f-9146-2a648932429d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.769427] env[63021]: DEBUG nova.network.neutron [req-4c418585-305b-45d0-984e-0f68035680fc req-b49f813e-035f-4674-8c7a-83ab6f3addb8 service nova] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Refreshing network info cache for port e581fd41-04b7-4081-a7be-c6389f4d623f {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 868.854944] env[63021]: DEBUG nova.network.neutron [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Updating instance_info_cache with network_info: [{"id": "37dcb3f1-2931-40c6-96ba-b31b8fa1a72b", "address": "fa:16:3e:57:13:17", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37dcb3f1-29", "ovs_interfaceid": "37dcb3f1-2931-40c6-96ba-b31b8fa1a72b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d92edf1f-d15d-4f56-9da3-92cbe44b6709", "address": "fa:16:3e:b5:1b:c8", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd92edf1f-d1", "ovs_interfaceid": "d92edf1f-d15d-4f56-9da3-92cbe44b6709", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.875733] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.933s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.880160] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.163s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.881691] env[63021]: INFO nova.compute.claims [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 868.906587] env[63021]: DEBUG oslo_vmware.api [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293804, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.917578] env[63021]: INFO nova.scheduler.client.report [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Deleted allocations for instance 36d26354-6127-4557-acb0-8743e27ff1ff [ 868.922049] env[63021]: DEBUG nova.network.neutron [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Updating instance_info_cache with network_info: [{"id": "3dc162c2-36e3-4bb0-b7a4-3116c11c145a", "address": "fa:16:3e:ab:61:b7", "network": {"id": "265e2e2e-6ec6-4d4c-8b4c-0a2e8baa6bdb", "bridge": "br-int", "label": "tempest-ImagesTestJSON-242691215-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7331caed42034dfc81c79cfabbe0829e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bfae3ef8-cae7-455d-8632-ba93e1671625", "external-id": "cl2-zone-841", "segmentation_id": 841, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3dc162c2-36", "ovs_interfaceid": "3dc162c2-36e3-4bb0-b7a4-3116c11c145a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.206436] env[63021]: INFO nova.compute.manager [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Took 37.16 seconds to build instance. [ 869.357202] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Releasing lock "refresh_cache-36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.357864] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.358030] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquired lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.360887] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb324f9d-35b2-4741-ad6a-6ab12cffb650 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.379440] env[63021]: DEBUG nova.virt.hardware [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 869.379725] env[63021]: DEBUG nova.virt.hardware [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 869.379832] env[63021]: DEBUG nova.virt.hardware [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 869.380010] env[63021]: DEBUG nova.virt.hardware [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 869.380213] env[63021]: DEBUG nova.virt.hardware [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 869.380386] env[63021]: DEBUG nova.virt.hardware [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 869.380603] env[63021]: DEBUG nova.virt.hardware [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 869.380763] env[63021]: DEBUG nova.virt.hardware [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 869.380923] env[63021]: DEBUG nova.virt.hardware [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 869.381111] env[63021]: DEBUG nova.virt.hardware [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 869.381285] env[63021]: DEBUG nova.virt.hardware [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 869.387612] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Reconfiguring VM to attach interface {{(pid=63021) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 869.391977] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0511f172-c994-4930-89a3-6ed2c15f1bcd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.415429] env[63021]: DEBUG oslo_vmware.api [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293804, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.93406} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.416669] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 16b21798-17e8-4aeb-affa-57eae31d5dd3/16b21798-17e8-4aeb-affa-57eae31d5dd3.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 869.416881] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 869.417195] env[63021]: DEBUG oslo_vmware.api [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 869.417195] env[63021]: value = "task-1293805" [ 869.417195] env[63021]: _type = "Task" [ 869.417195] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.417377] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4f4b9b44-365c-40f3-b769-93ab21a9d188 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.428805] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c4f6d92e-2e08-4279-b52c-0eb0f2a32e2a tempest-MultipleCreateTestJSON-1182003780 tempest-MultipleCreateTestJSON-1182003780-project-member] Lock "36d26354-6127-4557-acb0-8743e27ff1ff" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.512s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.428967] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Releasing lock "refresh_cache-881c2fb8-5f8c-48c2-a173-8fab949f24f7" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.429238] env[63021]: DEBUG nova.compute.manager [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Instance network_info: |[{"id": "3dc162c2-36e3-4bb0-b7a4-3116c11c145a", "address": "fa:16:3e:ab:61:b7", "network": {"id": "265e2e2e-6ec6-4d4c-8b4c-0a2e8baa6bdb", "bridge": "br-int", "label": "tempest-ImagesTestJSON-242691215-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7331caed42034dfc81c79cfabbe0829e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bfae3ef8-cae7-455d-8632-ba93e1671625", "external-id": "cl2-zone-841", "segmentation_id": 841, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3dc162c2-36", "ovs_interfaceid": "3dc162c2-36e3-4bb0-b7a4-3116c11c145a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 869.430611] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ab:61:b7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bfae3ef8-cae7-455d-8632-ba93e1671625', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3dc162c2-36e3-4bb0-b7a4-3116c11c145a', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 869.438735] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Creating folder: Project (7331caed42034dfc81c79cfabbe0829e). Parent ref: group-v277447. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 869.444317] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5c8611ba-cf12-4e9b-a01b-334500c6f58e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.446826] env[63021]: DEBUG oslo_vmware.api [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293805, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.446826] env[63021]: DEBUG oslo_vmware.api [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 869.446826] env[63021]: value = "task-1293806" [ 869.446826] env[63021]: _type = "Task" [ 869.446826] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.457919] env[63021]: DEBUG oslo_vmware.api [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293806, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.459457] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Created folder: Project (7331caed42034dfc81c79cfabbe0829e) in parent group-v277447. [ 869.459636] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Creating folder: Instances. Parent ref: group-v277522. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 869.459879] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-911104a2-b219-4432-8024-0646af38031a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.472092] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Created folder: Instances in parent group-v277522. [ 869.472092] env[63021]: DEBUG oslo.service.loopingcall [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 869.472092] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 869.474716] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7dab82a7-58cb-44a0-ac32-d982360222d0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.498793] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 869.498793] env[63021]: value = "task-1293809" [ 869.498793] env[63021]: _type = "Task" [ 869.498793] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.507717] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293809, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.530256] env[63021]: DEBUG nova.network.neutron [req-4c418585-305b-45d0-984e-0f68035680fc req-b49f813e-035f-4674-8c7a-83ab6f3addb8 service nova] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Updated VIF entry in instance network info cache for port e581fd41-04b7-4081-a7be-c6389f4d623f. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 869.530625] env[63021]: DEBUG nova.network.neutron [req-4c418585-305b-45d0-984e-0f68035680fc req-b49f813e-035f-4674-8c7a-83ab6f3addb8 service nova] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Updating instance_info_cache with network_info: [{"id": "e581fd41-04b7-4081-a7be-c6389f4d623f", "address": "fa:16:3e:5b:84:48", "network": {"id": "6d99496e-4e9f-4e5d-81a4-305f91e927cc", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-851847769-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "70e912b7ac9144c793323eaf26287ba1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f969bd9-e040-4b9b-85b2-7c61231584ad", "external-id": "nsx-vlan-transportzone-995", "segmentation_id": 995, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape581fd41-04", "ovs_interfaceid": "e581fd41-04b7-4081-a7be-c6389f4d623f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.559028] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Acquiring lock "ed26dff1-fed0-4baf-ad41-d14850254aec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.559028] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Lock "ed26dff1-fed0-4baf-ad41-d14850254aec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.708977] env[63021]: DEBUG oslo_concurrency.lockutils [None req-75dae2c9-b28f-4a49-8d0a-8a9487ed2bbb tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "c328d2f7-3398-4f25-b11c-f464be7af8a1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.559s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.930730] env[63021]: DEBUG oslo_vmware.api [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293805, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.960405] env[63021]: DEBUG oslo_vmware.api [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293806, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.114351} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.963368] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 869.966177] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9767e0d-24af-4f5d-8c59-17cb3729561d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.998585] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] 16b21798-17e8-4aeb-affa-57eae31d5dd3/16b21798-17e8-4aeb-affa-57eae31d5dd3.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 870.001442] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d8c12785-e9c2-4f30-b8c4-822e7e889650 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.030869] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293809, 'name': CreateVM_Task, 'duration_secs': 0.369072} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.035072] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 870.035632] env[63021]: DEBUG oslo_concurrency.lockutils [req-4c418585-305b-45d0-984e-0f68035680fc req-b49f813e-035f-4674-8c7a-83ab6f3addb8 service nova] Releasing lock "refresh_cache-7787f6bb-e855-476f-9146-2a648932429d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.036114] env[63021]: DEBUG oslo_vmware.api [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 870.036114] env[63021]: value = "task-1293810" [ 870.036114] env[63021]: _type = "Task" [ 870.036114] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.036986] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.037942] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.037942] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 870.037942] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c12bf862-5f8d-4814-8ca4-07296868d042 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.047014] env[63021]: DEBUG oslo_vmware.api [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 870.047014] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52c5fce3-ea68-84d4-c4f4-9315114d223c" [ 870.047014] env[63021]: _type = "Task" [ 870.047014] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.051601] env[63021]: DEBUG oslo_vmware.api [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293810, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.061504] env[63021]: DEBUG oslo_vmware.api [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52c5fce3-ea68-84d4-c4f4-9315114d223c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.216867] env[63021]: DEBUG nova.compute.manager [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 870.318526] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8290b014-13ab-440f-897d-3d1020f7b8b4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.327060] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3d67a35-a445-4c15-bd6f-ae033f5d86d2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.361496] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cc6305a-156e-43dc-9416-5aa5077cf88c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.370210] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c411fa74-8a7a-44f2-8473-983b99a4bdc3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.389842] env[63021]: DEBUG nova.compute.provider_tree [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 870.431973] env[63021]: DEBUG oslo_vmware.api [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293805, 'name': ReconfigVM_Task, 'duration_secs': 0.850143} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.432513] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Releasing lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.432719] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Reconfigured VM to attach interface {{(pid=63021) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 870.552126] env[63021]: DEBUG oslo_vmware.api [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293810, 'name': ReconfigVM_Task, 'duration_secs': 0.338412} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.552126] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Reconfigured VM instance instance-00000046 to attach disk [datastore2] 16b21798-17e8-4aeb-affa-57eae31d5dd3/16b21798-17e8-4aeb-affa-57eae31d5dd3.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 870.552126] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a52ae161-75cc-4e5a-84d5-dbfcf9654e20 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.565865] env[63021]: DEBUG oslo_vmware.api [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52c5fce3-ea68-84d4-c4f4-9315114d223c, 'name': SearchDatastore_Task, 'duration_secs': 0.013766} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.570636] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.570636] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 870.570636] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.570636] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.570636] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 870.570636] env[63021]: DEBUG oslo_vmware.api [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 870.570636] env[63021]: value = "task-1293811" [ 870.570636] env[63021]: _type = "Task" [ 870.570636] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.570636] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-93cc5c1f-8065-440a-8034-a6d1dc334f4a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.584564] env[63021]: DEBUG oslo_vmware.api [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293811, 'name': Rename_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.587120] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 870.588246] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 870.589446] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6ce25e8-b37c-4149-b3a6-aefbcc34170b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.597748] env[63021]: DEBUG oslo_vmware.api [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 870.597748] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52f0c31d-81cb-be1b-69be-4b5ffbb2a5a2" [ 870.597748] env[63021]: _type = "Task" [ 870.597748] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.609238] env[63021]: DEBUG oslo_vmware.api [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52f0c31d-81cb-be1b-69be-4b5ffbb2a5a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.740914] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.763653] env[63021]: DEBUG nova.compute.manager [req-aede1b15-2aec-4012-b030-0ced4caca34c req-e5604b53-77a1-475c-906f-cc38c812abd3 service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Received event network-vif-plugged-d92edf1f-d15d-4f56-9da3-92cbe44b6709 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 870.763861] env[63021]: DEBUG oslo_concurrency.lockutils [req-aede1b15-2aec-4012-b030-0ced4caca34c req-e5604b53-77a1-475c-906f-cc38c812abd3 service nova] Acquiring lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.764307] env[63021]: DEBUG oslo_concurrency.lockutils [req-aede1b15-2aec-4012-b030-0ced4caca34c req-e5604b53-77a1-475c-906f-cc38c812abd3 service nova] Lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.764536] env[63021]: DEBUG oslo_concurrency.lockutils [req-aede1b15-2aec-4012-b030-0ced4caca34c req-e5604b53-77a1-475c-906f-cc38c812abd3 service nova] Lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.764764] env[63021]: DEBUG nova.compute.manager [req-aede1b15-2aec-4012-b030-0ced4caca34c req-e5604b53-77a1-475c-906f-cc38c812abd3 service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] No waiting events found dispatching network-vif-plugged-d92edf1f-d15d-4f56-9da3-92cbe44b6709 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 870.764952] env[63021]: WARNING nova.compute.manager [req-aede1b15-2aec-4012-b030-0ced4caca34c req-e5604b53-77a1-475c-906f-cc38c812abd3 service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Received unexpected event network-vif-plugged-d92edf1f-d15d-4f56-9da3-92cbe44b6709 for instance with vm_state active and task_state None. [ 870.765130] env[63021]: DEBUG nova.compute.manager [req-aede1b15-2aec-4012-b030-0ced4caca34c req-e5604b53-77a1-475c-906f-cc38c812abd3 service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Received event network-changed-d92edf1f-d15d-4f56-9da3-92cbe44b6709 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 870.765757] env[63021]: DEBUG nova.compute.manager [req-aede1b15-2aec-4012-b030-0ced4caca34c req-e5604b53-77a1-475c-906f-cc38c812abd3 service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Refreshing instance network info cache due to event network-changed-d92edf1f-d15d-4f56-9da3-92cbe44b6709. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 870.765973] env[63021]: DEBUG oslo_concurrency.lockutils [req-aede1b15-2aec-4012-b030-0ced4caca34c req-e5604b53-77a1-475c-906f-cc38c812abd3 service nova] Acquiring lock "refresh_cache-36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.766104] env[63021]: DEBUG oslo_concurrency.lockutils [req-aede1b15-2aec-4012-b030-0ced4caca34c req-e5604b53-77a1-475c-906f-cc38c812abd3 service nova] Acquired lock "refresh_cache-36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.766270] env[63021]: DEBUG nova.network.neutron [req-aede1b15-2aec-4012-b030-0ced4caca34c req-e5604b53-77a1-475c-906f-cc38c812abd3 service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Refreshing network info cache for port d92edf1f-d15d-4f56-9da3-92cbe44b6709 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 870.896116] env[63021]: DEBUG nova.scheduler.client.report [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 870.937842] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ff233ce5-10cc-42d4-8a59-e9c19953cb4a tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "interface-36ec45ec-761b-4d62-a74f-e4d9a840ada0-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.880s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.961057] env[63021]: DEBUG nova.compute.manager [req-659a2f3c-5ce9-4b65-81f8-20d4e683933d req-7db12f11-b64d-49ff-bdcc-0ffa9195680f service nova] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Received event network-changed-e581fd41-04b7-4081-a7be-c6389f4d623f {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 870.961057] env[63021]: DEBUG nova.compute.manager [req-659a2f3c-5ce9-4b65-81f8-20d4e683933d req-7db12f11-b64d-49ff-bdcc-0ffa9195680f service nova] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Refreshing instance network info cache due to event network-changed-e581fd41-04b7-4081-a7be-c6389f4d623f. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 870.961057] env[63021]: DEBUG oslo_concurrency.lockutils [req-659a2f3c-5ce9-4b65-81f8-20d4e683933d req-7db12f11-b64d-49ff-bdcc-0ffa9195680f service nova] Acquiring lock "refresh_cache-7787f6bb-e855-476f-9146-2a648932429d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.961057] env[63021]: DEBUG oslo_concurrency.lockutils [req-659a2f3c-5ce9-4b65-81f8-20d4e683933d req-7db12f11-b64d-49ff-bdcc-0ffa9195680f service nova] Acquired lock "refresh_cache-7787f6bb-e855-476f-9146-2a648932429d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.961057] env[63021]: DEBUG nova.network.neutron [req-659a2f3c-5ce9-4b65-81f8-20d4e683933d req-7db12f11-b64d-49ff-bdcc-0ffa9195680f service nova] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Refreshing network info cache for port e581fd41-04b7-4081-a7be-c6389f4d623f {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 871.084312] env[63021]: DEBUG oslo_vmware.api [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293811, 'name': Rename_Task, 'duration_secs': 0.166716} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.084713] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 871.084990] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9719b53e-44b8-484e-bdd5-181adac3d54a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.093301] env[63021]: DEBUG oslo_vmware.api [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 871.093301] env[63021]: value = "task-1293812" [ 871.093301] env[63021]: _type = "Task" [ 871.093301] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.108994] env[63021]: DEBUG oslo_vmware.api [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293812, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.115224] env[63021]: DEBUG oslo_vmware.api [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52f0c31d-81cb-be1b-69be-4b5ffbb2a5a2, 'name': SearchDatastore_Task, 'duration_secs': 0.012823} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.116072] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66f15c9a-57d7-4145-ac91-ed1143972ab4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.122604] env[63021]: DEBUG oslo_vmware.api [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 871.122604] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]523a6e08-95ef-201e-1880-31094563509f" [ 871.122604] env[63021]: _type = "Task" [ 871.122604] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.132463] env[63021]: DEBUG oslo_vmware.api [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]523a6e08-95ef-201e-1880-31094563509f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.401096] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.521s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.401627] env[63021]: DEBUG nova.compute.manager [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 871.406605] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.653s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.410907] env[63021]: INFO nova.compute.claims [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 871.605646] env[63021]: DEBUG oslo_vmware.api [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293812, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.618130] env[63021]: DEBUG nova.network.neutron [req-aede1b15-2aec-4012-b030-0ced4caca34c req-e5604b53-77a1-475c-906f-cc38c812abd3 service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Updated VIF entry in instance network info cache for port d92edf1f-d15d-4f56-9da3-92cbe44b6709. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 871.618601] env[63021]: DEBUG nova.network.neutron [req-aede1b15-2aec-4012-b030-0ced4caca34c req-e5604b53-77a1-475c-906f-cc38c812abd3 service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Updating instance_info_cache with network_info: [{"id": "37dcb3f1-2931-40c6-96ba-b31b8fa1a72b", "address": "fa:16:3e:57:13:17", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37dcb3f1-29", "ovs_interfaceid": "37dcb3f1-2931-40c6-96ba-b31b8fa1a72b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d92edf1f-d15d-4f56-9da3-92cbe44b6709", "address": "fa:16:3e:b5:1b:c8", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd92edf1f-d1", "ovs_interfaceid": "d92edf1f-d15d-4f56-9da3-92cbe44b6709", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.636587] env[63021]: DEBUG oslo_vmware.api [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]523a6e08-95ef-201e-1880-31094563509f, 'name': SearchDatastore_Task, 'duration_secs': 0.013142} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.636776] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.637041] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 881c2fb8-5f8c-48c2-a173-8fab949f24f7/881c2fb8-5f8c-48c2-a173-8fab949f24f7.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 871.637304] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-12328041-9a21-4f4d-9b74-7386bd987bde {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.645527] env[63021]: DEBUG oslo_vmware.api [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 871.645527] env[63021]: value = "task-1293813" [ 871.645527] env[63021]: _type = "Task" [ 871.645527] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.655061] env[63021]: DEBUG oslo_vmware.api [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293813, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.769048] env[63021]: DEBUG nova.network.neutron [req-659a2f3c-5ce9-4b65-81f8-20d4e683933d req-7db12f11-b64d-49ff-bdcc-0ffa9195680f service nova] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Updated VIF entry in instance network info cache for port e581fd41-04b7-4081-a7be-c6389f4d623f. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 871.769504] env[63021]: DEBUG nova.network.neutron [req-659a2f3c-5ce9-4b65-81f8-20d4e683933d req-7db12f11-b64d-49ff-bdcc-0ffa9195680f service nova] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Updating instance_info_cache with network_info: [{"id": "e581fd41-04b7-4081-a7be-c6389f4d623f", "address": "fa:16:3e:5b:84:48", "network": {"id": "6d99496e-4e9f-4e5d-81a4-305f91e927cc", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-851847769-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "70e912b7ac9144c793323eaf26287ba1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f969bd9-e040-4b9b-85b2-7c61231584ad", "external-id": "nsx-vlan-transportzone-995", "segmentation_id": 995, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape581fd41-04", "ovs_interfaceid": "e581fd41-04b7-4081-a7be-c6389f4d623f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.913290] env[63021]: DEBUG nova.compute.utils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 871.918020] env[63021]: DEBUG nova.compute.manager [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 871.918020] env[63021]: DEBUG nova.network.neutron [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 871.969952] env[63021]: DEBUG nova.policy [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '305f29bd44c949afa5731ec54c1a296d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1ed12e42e778445eb8f937906ce5ae72', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 872.109353] env[63021]: DEBUG oslo_vmware.api [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293812, 'name': PowerOnVM_Task, 'duration_secs': 0.853045} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.114021] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 872.114021] env[63021]: INFO nova.compute.manager [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Took 9.60 seconds to spawn the instance on the hypervisor. [ 872.114021] env[63021]: DEBUG nova.compute.manager [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 872.114021] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce96cae4-d4fb-42a4-8862-135056b27b70 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.132590] env[63021]: DEBUG oslo_concurrency.lockutils [req-aede1b15-2aec-4012-b030-0ced4caca34c req-e5604b53-77a1-475c-906f-cc38c812abd3 service nova] Releasing lock "refresh_cache-36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.132590] env[63021]: DEBUG nova.compute.manager [req-aede1b15-2aec-4012-b030-0ced4caca34c req-e5604b53-77a1-475c-906f-cc38c812abd3 service nova] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Received event network-vif-plugged-3dc162c2-36e3-4bb0-b7a4-3116c11c145a {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 872.132590] env[63021]: DEBUG oslo_concurrency.lockutils [req-aede1b15-2aec-4012-b030-0ced4caca34c req-e5604b53-77a1-475c-906f-cc38c812abd3 service nova] Acquiring lock "881c2fb8-5f8c-48c2-a173-8fab949f24f7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.132590] env[63021]: DEBUG oslo_concurrency.lockutils [req-aede1b15-2aec-4012-b030-0ced4caca34c req-e5604b53-77a1-475c-906f-cc38c812abd3 service nova] Lock "881c2fb8-5f8c-48c2-a173-8fab949f24f7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.132590] env[63021]: DEBUG oslo_concurrency.lockutils [req-aede1b15-2aec-4012-b030-0ced4caca34c req-e5604b53-77a1-475c-906f-cc38c812abd3 service nova] Lock "881c2fb8-5f8c-48c2-a173-8fab949f24f7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.132590] env[63021]: DEBUG nova.compute.manager [req-aede1b15-2aec-4012-b030-0ced4caca34c req-e5604b53-77a1-475c-906f-cc38c812abd3 service nova] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] No waiting events found dispatching network-vif-plugged-3dc162c2-36e3-4bb0-b7a4-3116c11c145a {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 872.132590] env[63021]: WARNING nova.compute.manager [req-aede1b15-2aec-4012-b030-0ced4caca34c req-e5604b53-77a1-475c-906f-cc38c812abd3 service nova] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Received unexpected event network-vif-plugged-3dc162c2-36e3-4bb0-b7a4-3116c11c145a for instance with vm_state building and task_state spawning. [ 872.132590] env[63021]: DEBUG nova.compute.manager [req-aede1b15-2aec-4012-b030-0ced4caca34c req-e5604b53-77a1-475c-906f-cc38c812abd3 service nova] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Received event network-changed-3dc162c2-36e3-4bb0-b7a4-3116c11c145a {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 872.132590] env[63021]: DEBUG nova.compute.manager [req-aede1b15-2aec-4012-b030-0ced4caca34c req-e5604b53-77a1-475c-906f-cc38c812abd3 service nova] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Refreshing instance network info cache due to event network-changed-3dc162c2-36e3-4bb0-b7a4-3116c11c145a. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 872.132590] env[63021]: DEBUG oslo_concurrency.lockutils [req-aede1b15-2aec-4012-b030-0ced4caca34c req-e5604b53-77a1-475c-906f-cc38c812abd3 service nova] Acquiring lock "refresh_cache-881c2fb8-5f8c-48c2-a173-8fab949f24f7" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.132590] env[63021]: DEBUG oslo_concurrency.lockutils [req-aede1b15-2aec-4012-b030-0ced4caca34c req-e5604b53-77a1-475c-906f-cc38c812abd3 service nova] Acquired lock "refresh_cache-881c2fb8-5f8c-48c2-a173-8fab949f24f7" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.132590] env[63021]: DEBUG nova.network.neutron [req-aede1b15-2aec-4012-b030-0ced4caca34c req-e5604b53-77a1-475c-906f-cc38c812abd3 service nova] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Refreshing network info cache for port 3dc162c2-36e3-4bb0-b7a4-3116c11c145a {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 872.157281] env[63021]: DEBUG oslo_vmware.api [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293813, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.276837] env[63021]: DEBUG oslo_concurrency.lockutils [req-659a2f3c-5ce9-4b65-81f8-20d4e683933d req-7db12f11-b64d-49ff-bdcc-0ffa9195680f service nova] Releasing lock "refresh_cache-7787f6bb-e855-476f-9146-2a648932429d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.276837] env[63021]: DEBUG nova.compute.manager [req-659a2f3c-5ce9-4b65-81f8-20d4e683933d req-7db12f11-b64d-49ff-bdcc-0ffa9195680f service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Received event network-changed-43e48c53-d06d-41ff-97dd-a7747d5551d9 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 872.276837] env[63021]: DEBUG nova.compute.manager [req-659a2f3c-5ce9-4b65-81f8-20d4e683933d req-7db12f11-b64d-49ff-bdcc-0ffa9195680f service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Refreshing instance network info cache due to event network-changed-43e48c53-d06d-41ff-97dd-a7747d5551d9. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 872.276837] env[63021]: DEBUG oslo_concurrency.lockutils [req-659a2f3c-5ce9-4b65-81f8-20d4e683933d req-7db12f11-b64d-49ff-bdcc-0ffa9195680f service nova] Acquiring lock "refresh_cache-45c9e2be-eb19-4b83-b280-c9eeaddfccba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.276837] env[63021]: DEBUG oslo_concurrency.lockutils [req-659a2f3c-5ce9-4b65-81f8-20d4e683933d req-7db12f11-b64d-49ff-bdcc-0ffa9195680f service nova] Acquired lock "refresh_cache-45c9e2be-eb19-4b83-b280-c9eeaddfccba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.276837] env[63021]: DEBUG nova.network.neutron [req-659a2f3c-5ce9-4b65-81f8-20d4e683933d req-7db12f11-b64d-49ff-bdcc-0ffa9195680f service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Refreshing network info cache for port 43e48c53-d06d-41ff-97dd-a7747d5551d9 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 872.345939] env[63021]: DEBUG nova.network.neutron [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Successfully created port: ce58eb3e-a5f1-4d1a-ab2a-dd6fcb0d4f62 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 872.424275] env[63021]: DEBUG nova.compute.manager [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 872.633992] env[63021]: INFO nova.compute.manager [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Took 40.08 seconds to build instance. [ 872.660133] env[63021]: DEBUG oslo_vmware.api [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293813, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.770951} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.660413] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 881c2fb8-5f8c-48c2-a173-8fab949f24f7/881c2fb8-5f8c-48c2-a173-8fab949f24f7.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 872.660639] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 872.660887] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4c3d97fb-6897-4d1c-a6a0-e400dbf5ed93 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.670146] env[63021]: DEBUG oslo_vmware.api [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 872.670146] env[63021]: value = "task-1293814" [ 872.670146] env[63021]: _type = "Task" [ 872.670146] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.683620] env[63021]: DEBUG oslo_vmware.api [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293814, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.805681] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cdf9062-193a-4e73-9c68-1d32adf339f9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.813748] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b6105db-aaef-435b-a818-b7a64d651adc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.847125] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0ad36b1-969e-4081-ab97-20e889909f27 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.856023] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65e09731-bc4d-4cb2-aa2b-2dc6895663b1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.876503] env[63021]: DEBUG nova.compute.provider_tree [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 872.915894] env[63021]: DEBUG nova.network.neutron [req-aede1b15-2aec-4012-b030-0ced4caca34c req-e5604b53-77a1-475c-906f-cc38c812abd3 service nova] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Updated VIF entry in instance network info cache for port 3dc162c2-36e3-4bb0-b7a4-3116c11c145a. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 872.916472] env[63021]: DEBUG nova.network.neutron [req-aede1b15-2aec-4012-b030-0ced4caca34c req-e5604b53-77a1-475c-906f-cc38c812abd3 service nova] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Updating instance_info_cache with network_info: [{"id": "3dc162c2-36e3-4bb0-b7a4-3116c11c145a", "address": "fa:16:3e:ab:61:b7", "network": {"id": "265e2e2e-6ec6-4d4c-8b4c-0a2e8baa6bdb", "bridge": "br-int", "label": "tempest-ImagesTestJSON-242691215-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7331caed42034dfc81c79cfabbe0829e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bfae3ef8-cae7-455d-8632-ba93e1671625", "external-id": "cl2-zone-841", "segmentation_id": 841, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3dc162c2-36", "ovs_interfaceid": "3dc162c2-36e3-4bb0-b7a4-3116c11c145a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.084594] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Acquiring lock "7787f6bb-e855-476f-9146-2a648932429d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.084742] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Lock "7787f6bb-e855-476f-9146-2a648932429d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.084955] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Acquiring lock "7787f6bb-e855-476f-9146-2a648932429d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.085177] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Lock "7787f6bb-e855-476f-9146-2a648932429d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.085334] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Lock "7787f6bb-e855-476f-9146-2a648932429d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.087600] env[63021]: INFO nova.compute.manager [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Terminating instance [ 873.089479] env[63021]: DEBUG nova.compute.manager [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 873.089760] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 873.090602] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac0f2699-7e5c-491e-9fb7-3372c57ad835 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.099520] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 873.099838] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3432a02b-fad7-4463-b9c2-84742dc4359c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.108113] env[63021]: DEBUG oslo_vmware.api [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Waiting for the task: (returnval){ [ 873.108113] env[63021]: value = "task-1293815" [ 873.108113] env[63021]: _type = "Task" [ 873.108113] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.118900] env[63021]: DEBUG oslo_vmware.api [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': task-1293815, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.140617] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6e8631ca-80d5-479d-9405-5b05ec10ddd0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "16b21798-17e8-4aeb-affa-57eae31d5dd3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.487s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.182686] env[63021]: DEBUG oslo_vmware.api [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293814, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.140974} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.183078] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 873.183917] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-407ffa40-9555-422b-8577-f833ba0a3926 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.216658] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] 881c2fb8-5f8c-48c2-a173-8fab949f24f7/881c2fb8-5f8c-48c2-a173-8fab949f24f7.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 873.220378] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-75435c66-92f2-4ce1-8bb3-00b2d4f998b9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.247552] env[63021]: DEBUG oslo_vmware.api [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 873.247552] env[63021]: value = "task-1293816" [ 873.247552] env[63021]: _type = "Task" [ 873.247552] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.256945] env[63021]: DEBUG oslo_vmware.api [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293816, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.315955] env[63021]: DEBUG nova.compute.manager [req-b67a5d58-c765-46b9-9275-b1e88aea5ee7 req-551d6b95-f0cb-4ac4-90fe-82fc91eb5efb service nova] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Received event network-changed-eff6f35a-0a13-4e0c-88b0-96b66b240f9b {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 873.316217] env[63021]: DEBUG nova.compute.manager [req-b67a5d58-c765-46b9-9275-b1e88aea5ee7 req-551d6b95-f0cb-4ac4-90fe-82fc91eb5efb service nova] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Refreshing instance network info cache due to event network-changed-eff6f35a-0a13-4e0c-88b0-96b66b240f9b. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 873.316385] env[63021]: DEBUG oslo_concurrency.lockutils [req-b67a5d58-c765-46b9-9275-b1e88aea5ee7 req-551d6b95-f0cb-4ac4-90fe-82fc91eb5efb service nova] Acquiring lock "refresh_cache-c328d2f7-3398-4f25-b11c-f464be7af8a1" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.316565] env[63021]: DEBUG oslo_concurrency.lockutils [req-b67a5d58-c765-46b9-9275-b1e88aea5ee7 req-551d6b95-f0cb-4ac4-90fe-82fc91eb5efb service nova] Acquired lock "refresh_cache-c328d2f7-3398-4f25-b11c-f464be7af8a1" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.316667] env[63021]: DEBUG nova.network.neutron [req-b67a5d58-c765-46b9-9275-b1e88aea5ee7 req-551d6b95-f0cb-4ac4-90fe-82fc91eb5efb service nova] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Refreshing network info cache for port eff6f35a-0a13-4e0c-88b0-96b66b240f9b {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 873.379523] env[63021]: DEBUG nova.scheduler.client.report [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 873.419367] env[63021]: DEBUG oslo_concurrency.lockutils [req-aede1b15-2aec-4012-b030-0ced4caca34c req-e5604b53-77a1-475c-906f-cc38c812abd3 service nova] Releasing lock "refresh_cache-881c2fb8-5f8c-48c2-a173-8fab949f24f7" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.435897] env[63021]: DEBUG nova.compute.manager [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 873.466488] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 873.466732] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 873.466886] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 873.467072] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 873.467217] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 873.467357] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 873.467584] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 873.467760] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 873.467924] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 873.468093] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 873.468263] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 873.469166] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ae8e325-efde-4dff-a38a-6696e43cf0c9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.477952] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e9dd659-c05a-4b84-b730-0a4af5e1df8b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.482226] env[63021]: DEBUG nova.network.neutron [req-659a2f3c-5ce9-4b65-81f8-20d4e683933d req-7db12f11-b64d-49ff-bdcc-0ffa9195680f service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Updated VIF entry in instance network info cache for port 43e48c53-d06d-41ff-97dd-a7747d5551d9. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 873.482553] env[63021]: DEBUG nova.network.neutron [req-659a2f3c-5ce9-4b65-81f8-20d4e683933d req-7db12f11-b64d-49ff-bdcc-0ffa9195680f service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Updating instance_info_cache with network_info: [{"id": "43e48c53-d06d-41ff-97dd-a7747d5551d9", "address": "fa:16:3e:73:e1:9e", "network": {"id": "36548b96-bd4d-4725-8dab-5cb1cca12c4e", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-672805719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.178", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b4c8cd929ff4068b45ee829b19f9769", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8614db14-cc04-466b-b309-367ab8296cda", "external-id": "nsx-vlan-transportzone-337", "segmentation_id": 337, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43e48c53-d0", "ovs_interfaceid": "43e48c53-d06d-41ff-97dd-a7747d5551d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.620150] env[63021]: DEBUG oslo_vmware.api [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': task-1293815, 'name': PowerOffVM_Task, 'duration_secs': 0.268192} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.620441] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 873.620610] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 873.620867] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7127d413-ee36-4a02-85ad-0c61d4b995ed {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.722611] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 873.722950] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 873.723075] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Deleting the datastore file [datastore1] 7787f6bb-e855-476f-9146-2a648932429d {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 873.723303] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-80b9b225-d71a-4378-8b2f-a4d66f805ab1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.730952] env[63021]: DEBUG oslo_vmware.api [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Waiting for the task: (returnval){ [ 873.730952] env[63021]: value = "task-1293818" [ 873.730952] env[63021]: _type = "Task" [ 873.730952] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.739792] env[63021]: DEBUG oslo_vmware.api [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': task-1293818, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.757646] env[63021]: DEBUG oslo_vmware.api [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293816, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.888892] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.484s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.889467] env[63021]: DEBUG nova.compute.manager [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 873.893040] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.022s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.893442] env[63021]: DEBUG nova.objects.instance [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Lazy-loading 'resources' on Instance uuid 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 873.953612] env[63021]: DEBUG nova.network.neutron [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Successfully updated port: ce58eb3e-a5f1-4d1a-ab2a-dd6fcb0d4f62 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 873.985136] env[63021]: DEBUG oslo_concurrency.lockutils [req-659a2f3c-5ce9-4b65-81f8-20d4e683933d req-7db12f11-b64d-49ff-bdcc-0ffa9195680f service nova] Releasing lock "refresh_cache-45c9e2be-eb19-4b83-b280-c9eeaddfccba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.985590] env[63021]: DEBUG nova.compute.manager [req-659a2f3c-5ce9-4b65-81f8-20d4e683933d req-7db12f11-b64d-49ff-bdcc-0ffa9195680f service nova] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Received event network-changed-e581fd41-04b7-4081-a7be-c6389f4d623f {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 873.985912] env[63021]: DEBUG nova.compute.manager [req-659a2f3c-5ce9-4b65-81f8-20d4e683933d req-7db12f11-b64d-49ff-bdcc-0ffa9195680f service nova] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Refreshing instance network info cache due to event network-changed-e581fd41-04b7-4081-a7be-c6389f4d623f. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 873.986187] env[63021]: DEBUG oslo_concurrency.lockutils [req-659a2f3c-5ce9-4b65-81f8-20d4e683933d req-7db12f11-b64d-49ff-bdcc-0ffa9195680f service nova] Acquiring lock "refresh_cache-7787f6bb-e855-476f-9146-2a648932429d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.986338] env[63021]: DEBUG oslo_concurrency.lockutils [req-659a2f3c-5ce9-4b65-81f8-20d4e683933d req-7db12f11-b64d-49ff-bdcc-0ffa9195680f service nova] Acquired lock "refresh_cache-7787f6bb-e855-476f-9146-2a648932429d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.986516] env[63021]: DEBUG nova.network.neutron [req-659a2f3c-5ce9-4b65-81f8-20d4e683933d req-7db12f11-b64d-49ff-bdcc-0ffa9195680f service nova] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Refreshing network info cache for port e581fd41-04b7-4081-a7be-c6389f4d623f {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 874.081139] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "16b21798-17e8-4aeb-affa-57eae31d5dd3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.081400] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "16b21798-17e8-4aeb-affa-57eae31d5dd3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.081662] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "16b21798-17e8-4aeb-affa-57eae31d5dd3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.081916] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "16b21798-17e8-4aeb-affa-57eae31d5dd3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.082170] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "16b21798-17e8-4aeb-affa-57eae31d5dd3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.084601] env[63021]: INFO nova.compute.manager [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Terminating instance [ 874.086552] env[63021]: DEBUG nova.compute.manager [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 874.086754] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 874.087667] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd587c60-b8aa-48b7-b5ec-ab570badbb35 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.096767] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 874.097140] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8f6e3c65-f3b4-4eac-bfd5-5596efaf06c9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.105539] env[63021]: DEBUG oslo_vmware.api [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 874.105539] env[63021]: value = "task-1293819" [ 874.105539] env[63021]: _type = "Task" [ 874.105539] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.109739] env[63021]: DEBUG nova.network.neutron [req-b67a5d58-c765-46b9-9275-b1e88aea5ee7 req-551d6b95-f0cb-4ac4-90fe-82fc91eb5efb service nova] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Updated VIF entry in instance network info cache for port eff6f35a-0a13-4e0c-88b0-96b66b240f9b. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 874.110110] env[63021]: DEBUG nova.network.neutron [req-b67a5d58-c765-46b9-9275-b1e88aea5ee7 req-551d6b95-f0cb-4ac4-90fe-82fc91eb5efb service nova] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Updating instance_info_cache with network_info: [{"id": "eff6f35a-0a13-4e0c-88b0-96b66b240f9b", "address": "fa:16:3e:27:3b:5e", "network": {"id": "9502e364-60e1-47ef-a1f7-f3e3ff45cf69", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-164717784-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5af87713ae6b477ca8a584aed287889d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeff6f35a-0a", "ovs_interfaceid": "eff6f35a-0a13-4e0c-88b0-96b66b240f9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.117938] env[63021]: DEBUG oslo_vmware.api [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293819, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.120624] env[63021]: DEBUG nova.compute.manager [req-9787e6b9-f0f6-429e-b810-a5e5e32af2ac req-2b6e52f4-edc4-4500-b5b4-6b1bce3d5985 service nova] [instance: f0154308-e6a8-4321-a082-99291344664e] Received event network-vif-plugged-ce58eb3e-a5f1-4d1a-ab2a-dd6fcb0d4f62 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 874.120844] env[63021]: DEBUG oslo_concurrency.lockutils [req-9787e6b9-f0f6-429e-b810-a5e5e32af2ac req-2b6e52f4-edc4-4500-b5b4-6b1bce3d5985 service nova] Acquiring lock "f0154308-e6a8-4321-a082-99291344664e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.121158] env[63021]: DEBUG oslo_concurrency.lockutils [req-9787e6b9-f0f6-429e-b810-a5e5e32af2ac req-2b6e52f4-edc4-4500-b5b4-6b1bce3d5985 service nova] Lock "f0154308-e6a8-4321-a082-99291344664e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.121420] env[63021]: DEBUG oslo_concurrency.lockutils [req-9787e6b9-f0f6-429e-b810-a5e5e32af2ac req-2b6e52f4-edc4-4500-b5b4-6b1bce3d5985 service nova] Lock "f0154308-e6a8-4321-a082-99291344664e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.121691] env[63021]: DEBUG nova.compute.manager [req-9787e6b9-f0f6-429e-b810-a5e5e32af2ac req-2b6e52f4-edc4-4500-b5b4-6b1bce3d5985 service nova] [instance: f0154308-e6a8-4321-a082-99291344664e] No waiting events found dispatching network-vif-plugged-ce58eb3e-a5f1-4d1a-ab2a-dd6fcb0d4f62 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 874.121881] env[63021]: WARNING nova.compute.manager [req-9787e6b9-f0f6-429e-b810-a5e5e32af2ac req-2b6e52f4-edc4-4500-b5b4-6b1bce3d5985 service nova] [instance: f0154308-e6a8-4321-a082-99291344664e] Received unexpected event network-vif-plugged-ce58eb3e-a5f1-4d1a-ab2a-dd6fcb0d4f62 for instance with vm_state building and task_state spawning. [ 874.185509] env[63021]: DEBUG oslo_concurrency.lockutils [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "interface-36ec45ec-761b-4d62-a74f-e4d9a840ada0-2f88c221-8f47-478e-a5f3-ba0a230c7ad4" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.185984] env[63021]: DEBUG oslo_concurrency.lockutils [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "interface-36ec45ec-761b-4d62-a74f-e4d9a840ada0-2f88c221-8f47-478e-a5f3-ba0a230c7ad4" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.186639] env[63021]: DEBUG nova.objects.instance [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lazy-loading 'flavor' on Instance uuid 36ec45ec-761b-4d62-a74f-e4d9a840ada0 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 874.241321] env[63021]: DEBUG oslo_vmware.api [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': task-1293818, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.260087] env[63021]: DEBUG oslo_vmware.api [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293816, 'name': ReconfigVM_Task, 'duration_secs': 0.961021} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.260389] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Reconfigured VM instance instance-00000047 to attach disk [datastore2] 881c2fb8-5f8c-48c2-a173-8fab949f24f7/881c2fb8-5f8c-48c2-a173-8fab949f24f7.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 874.261039] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d9416f5f-8b0a-41c6-a40c-33a54a3dc527 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.268413] env[63021]: DEBUG oslo_vmware.api [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 874.268413] env[63021]: value = "task-1293820" [ 874.268413] env[63021]: _type = "Task" [ 874.268413] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.277021] env[63021]: DEBUG oslo_vmware.api [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293820, 'name': Rename_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.397586] env[63021]: DEBUG nova.compute.utils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 874.399067] env[63021]: DEBUG nova.compute.manager [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 874.399241] env[63021]: DEBUG nova.network.neutron [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 874.453120] env[63021]: DEBUG nova.policy [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '305f29bd44c949afa5731ec54c1a296d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1ed12e42e778445eb8f937906ce5ae72', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 874.457571] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquiring lock "refresh_cache-f0154308-e6a8-4321-a082-99291344664e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.457571] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquired lock "refresh_cache-f0154308-e6a8-4321-a082-99291344664e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.457571] env[63021]: DEBUG nova.network.neutron [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 874.615372] env[63021]: DEBUG oslo_concurrency.lockutils [req-b67a5d58-c765-46b9-9275-b1e88aea5ee7 req-551d6b95-f0cb-4ac4-90fe-82fc91eb5efb service nova] Releasing lock "refresh_cache-c328d2f7-3398-4f25-b11c-f464be7af8a1" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.622041] env[63021]: DEBUG oslo_vmware.api [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293819, 'name': PowerOffVM_Task, 'duration_secs': 0.240362} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.624737] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 874.624914] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 874.625350] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b1e19f74-6ff0-450d-af19-b7db1bcf4d34 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.688521] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b84062bc-b27d-4811-a27b-3a673a7231aa {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.694682] env[63021]: DEBUG nova.network.neutron [req-659a2f3c-5ce9-4b65-81f8-20d4e683933d req-7db12f11-b64d-49ff-bdcc-0ffa9195680f service nova] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Updated VIF entry in instance network info cache for port e581fd41-04b7-4081-a7be-c6389f4d623f. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 874.697575] env[63021]: DEBUG nova.network.neutron [req-659a2f3c-5ce9-4b65-81f8-20d4e683933d req-7db12f11-b64d-49ff-bdcc-0ffa9195680f service nova] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Updating instance_info_cache with network_info: [{"id": "e581fd41-04b7-4081-a7be-c6389f4d623f", "address": "fa:16:3e:5b:84:48", "network": {"id": "6d99496e-4e9f-4e5d-81a4-305f91e927cc", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-851847769-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "70e912b7ac9144c793323eaf26287ba1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f969bd9-e040-4b9b-85b2-7c61231584ad", "external-id": "nsx-vlan-transportzone-995", "segmentation_id": 995, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape581fd41-04", "ovs_interfaceid": "e581fd41-04b7-4081-a7be-c6389f4d623f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.708008] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e564653-7b43-4d6e-ba14-ac469e16ccce {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.714831] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 874.715243] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Deleting contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 874.715443] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Deleting the datastore file [datastore2] 16b21798-17e8-4aeb-affa-57eae31d5dd3 {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 874.741035] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fd9978c0-ff9b-4fd5-a370-89e0583361b8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.749475] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c26d7efe-2411-463e-a1c6-3268a88efe73 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.768317] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-308850d1-19c2-46bb-9572-a84082c4aa10 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.769827] env[63021]: DEBUG oslo_vmware.api [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 874.769827] env[63021]: value = "task-1293822" [ 874.769827] env[63021]: _type = "Task" [ 874.769827] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.770123] env[63021]: DEBUG oslo_vmware.api [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Task: {'id': task-1293818, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.652699} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.770712] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 874.770895] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 874.771469] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 874.771469] env[63021]: INFO nova.compute.manager [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Took 1.68 seconds to destroy the instance on the hypervisor. [ 874.771579] env[63021]: DEBUG oslo.service.loopingcall [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 874.783023] env[63021]: DEBUG nova.compute.manager [-] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 874.783023] env[63021]: DEBUG nova.network.neutron [-] [instance: 7787f6bb-e855-476f-9146-2a648932429d] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 874.788958] env[63021]: DEBUG nova.compute.provider_tree [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 874.802473] env[63021]: DEBUG oslo_vmware.api [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293822, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.803156] env[63021]: DEBUG nova.network.neutron [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Successfully created port: fb56a4a1-033a-4205-aced-43f632d2f594 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 874.808207] env[63021]: DEBUG oslo_vmware.api [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293820, 'name': Rename_Task, 'duration_secs': 0.18627} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.808801] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 874.809105] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c729f10b-4019-4df0-9cdb-8cee99378d21 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.817140] env[63021]: DEBUG oslo_vmware.api [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 874.817140] env[63021]: value = "task-1293823" [ 874.817140] env[63021]: _type = "Task" [ 874.817140] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.826251] env[63021]: DEBUG oslo_vmware.api [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293823, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.872606] env[63021]: DEBUG nova.objects.instance [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lazy-loading 'pci_requests' on Instance uuid 36ec45ec-761b-4d62-a74f-e4d9a840ada0 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 874.904263] env[63021]: DEBUG nova.compute.manager [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 875.013147] env[63021]: DEBUG nova.network.neutron [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 875.202735] env[63021]: DEBUG oslo_concurrency.lockutils [req-659a2f3c-5ce9-4b65-81f8-20d4e683933d req-7db12f11-b64d-49ff-bdcc-0ffa9195680f service nova] Releasing lock "refresh_cache-7787f6bb-e855-476f-9146-2a648932429d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.283126] env[63021]: DEBUG oslo_vmware.api [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293822, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.268687} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.283453] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 875.283644] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Deleted contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 875.283822] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 875.283997] env[63021]: INFO nova.compute.manager [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Took 1.20 seconds to destroy the instance on the hypervisor. [ 875.284255] env[63021]: DEBUG oslo.service.loopingcall [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 875.284447] env[63021]: DEBUG nova.compute.manager [-] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 875.284777] env[63021]: DEBUG nova.network.neutron [-] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 875.297946] env[63021]: DEBUG nova.scheduler.client.report [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 875.331286] env[63021]: DEBUG oslo_vmware.api [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293823, 'name': PowerOnVM_Task, 'duration_secs': 0.48536} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.331554] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 875.331949] env[63021]: INFO nova.compute.manager [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Took 8.36 seconds to spawn the instance on the hypervisor. [ 875.332152] env[63021]: DEBUG nova.compute.manager [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 875.333188] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb344bf2-07f1-4879-9066-7859ac643e10 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.375505] env[63021]: DEBUG nova.objects.base [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Object Instance<36ec45ec-761b-4d62-a74f-e4d9a840ada0> lazy-loaded attributes: flavor,pci_requests {{(pid=63021) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 875.375738] env[63021]: DEBUG nova.network.neutron [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 875.423230] env[63021]: DEBUG nova.network.neutron [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Updating instance_info_cache with network_info: [{"id": "ce58eb3e-a5f1-4d1a-ab2a-dd6fcb0d4f62", "address": "fa:16:3e:12:e0:99", "network": {"id": "3e88ae40-5e21-4cbb-a0b4-84e0bd2910fa", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-557141593-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ed12e42e778445eb8f937906ce5ae72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce58eb3e-a5", "ovs_interfaceid": "ce58eb3e-a5f1-4d1a-ab2a-dd6fcb0d4f62", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.504431] env[63021]: DEBUG nova.policy [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6ea19453dbdb4251a93dde2576aeff64', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '58693d1e991249f28914a39761cc2627', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 875.804868] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.912s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.811229] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.596s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.812814] env[63021]: INFO nova.compute.claims [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 875.843314] env[63021]: INFO nova.scheduler.client.report [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Deleted allocations for instance 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4 [ 875.854427] env[63021]: DEBUG nova.network.neutron [-] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.857698] env[63021]: INFO nova.compute.manager [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Took 40.22 seconds to build instance. [ 875.913453] env[63021]: DEBUG nova.compute.manager [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 875.926373] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Releasing lock "refresh_cache-f0154308-e6a8-4321-a082-99291344664e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.926373] env[63021]: DEBUG nova.compute.manager [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Instance network_info: |[{"id": "ce58eb3e-a5f1-4d1a-ab2a-dd6fcb0d4f62", "address": "fa:16:3e:12:e0:99", "network": {"id": "3e88ae40-5e21-4cbb-a0b4-84e0bd2910fa", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-557141593-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ed12e42e778445eb8f937906ce5ae72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce58eb3e-a5", "ovs_interfaceid": "ce58eb3e-a5f1-4d1a-ab2a-dd6fcb0d4f62", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 875.926373] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:12:e0:99', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b1f3e6c3-5584-4852-9017-476ab8ac4946', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ce58eb3e-a5f1-4d1a-ab2a-dd6fcb0d4f62', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 875.933561] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Creating folder: Project (1ed12e42e778445eb8f937906ce5ae72). Parent ref: group-v277447. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 875.933884] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-39b78f40-fbaa-4d08-8ba2-184dc8b365b1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.949258] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 875.949258] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 875.949258] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 875.949258] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 875.949258] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 875.949258] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 875.949258] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 875.949774] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 875.950133] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 875.950436] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 875.950727] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 875.951722] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab6fd234-7589-4907-8053-83e56fb0f3d5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.956138] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Created folder: Project (1ed12e42e778445eb8f937906ce5ae72) in parent group-v277447. [ 875.956869] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Creating folder: Instances. Parent ref: group-v277525. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 875.957921] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2203688d-4e4c-4414-a87b-fdf49ce7e74c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.963229] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b7a52e8-ce0d-4300-98a9-137eea4c7b35 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.970514] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Created folder: Instances in parent group-v277525. [ 875.970767] env[63021]: DEBUG oslo.service.loopingcall [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 875.978526] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f0154308-e6a8-4321-a082-99291344664e] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 875.980024] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-79783372-41f0-4fd9-ab7e-51b4b2673acb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.000186] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 876.000186] env[63021]: value = "task-1293826" [ 876.000186] env[63021]: _type = "Task" [ 876.000186] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.010923] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293826, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.236645] env[63021]: DEBUG nova.network.neutron [-] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.349742] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ac4a1ad2-814b-4524-94ba-2315bcc0f86f tempest-ServersNegativeTestMultiTenantJSON-1521458590 tempest-ServersNegativeTestMultiTenantJSON-1521458590-project-member] Lock "2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.513s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.358382] env[63021]: INFO nova.compute.manager [-] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Took 1.58 seconds to deallocate network for instance. [ 876.360295] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d80d1b7e-9377-431f-a394-9a836447aaee tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "881c2fb8-5f8c-48c2-a173-8fab949f24f7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.397s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.510808] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293826, 'name': CreateVM_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.630227] env[63021]: DEBUG nova.compute.manager [req-b2dc0cc9-a998-4967-909a-f7ff60fb5069 req-f68bd29d-0e1c-491e-a74a-e00cc06fa78d service nova] [instance: f0154308-e6a8-4321-a082-99291344664e] Received event network-changed-ce58eb3e-a5f1-4d1a-ab2a-dd6fcb0d4f62 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 876.630428] env[63021]: DEBUG nova.compute.manager [req-b2dc0cc9-a998-4967-909a-f7ff60fb5069 req-f68bd29d-0e1c-491e-a74a-e00cc06fa78d service nova] [instance: f0154308-e6a8-4321-a082-99291344664e] Refreshing instance network info cache due to event network-changed-ce58eb3e-a5f1-4d1a-ab2a-dd6fcb0d4f62. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 876.630640] env[63021]: DEBUG oslo_concurrency.lockutils [req-b2dc0cc9-a998-4967-909a-f7ff60fb5069 req-f68bd29d-0e1c-491e-a74a-e00cc06fa78d service nova] Acquiring lock "refresh_cache-f0154308-e6a8-4321-a082-99291344664e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.630771] env[63021]: DEBUG oslo_concurrency.lockutils [req-b2dc0cc9-a998-4967-909a-f7ff60fb5069 req-f68bd29d-0e1c-491e-a74a-e00cc06fa78d service nova] Acquired lock "refresh_cache-f0154308-e6a8-4321-a082-99291344664e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.630927] env[63021]: DEBUG nova.network.neutron [req-b2dc0cc9-a998-4967-909a-f7ff60fb5069 req-f68bd29d-0e1c-491e-a74a-e00cc06fa78d service nova] [instance: f0154308-e6a8-4321-a082-99291344664e] Refreshing network info cache for port ce58eb3e-a5f1-4d1a-ab2a-dd6fcb0d4f62 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 876.740306] env[63021]: INFO nova.compute.manager [-] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Took 1.45 seconds to deallocate network for instance. [ 876.867502] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.870933] env[63021]: DEBUG nova.compute.manager [req-92bfdc22-265d-4767-9f2d-b568c5a65f11 req-d5bb1655-4763-4734-b68c-dfa9889266e0 service nova] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Received event network-vif-plugged-fb56a4a1-033a-4205-aced-43f632d2f594 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 876.871160] env[63021]: DEBUG oslo_concurrency.lockutils [req-92bfdc22-265d-4767-9f2d-b568c5a65f11 req-d5bb1655-4763-4734-b68c-dfa9889266e0 service nova] Acquiring lock "44dadf8e-a727-448f-887d-6408f7ebfbc6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.871367] env[63021]: DEBUG oslo_concurrency.lockutils [req-92bfdc22-265d-4767-9f2d-b568c5a65f11 req-d5bb1655-4763-4734-b68c-dfa9889266e0 service nova] Lock "44dadf8e-a727-448f-887d-6408f7ebfbc6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.871530] env[63021]: DEBUG oslo_concurrency.lockutils [req-92bfdc22-265d-4767-9f2d-b568c5a65f11 req-d5bb1655-4763-4734-b68c-dfa9889266e0 service nova] Lock "44dadf8e-a727-448f-887d-6408f7ebfbc6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.871695] env[63021]: DEBUG nova.compute.manager [req-92bfdc22-265d-4767-9f2d-b568c5a65f11 req-d5bb1655-4763-4734-b68c-dfa9889266e0 service nova] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] No waiting events found dispatching network-vif-plugged-fb56a4a1-033a-4205-aced-43f632d2f594 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 876.871853] env[63021]: WARNING nova.compute.manager [req-92bfdc22-265d-4767-9f2d-b568c5a65f11 req-d5bb1655-4763-4734-b68c-dfa9889266e0 service nova] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Received unexpected event network-vif-plugged-fb56a4a1-033a-4205-aced-43f632d2f594 for instance with vm_state building and task_state spawning. [ 876.920164] env[63021]: DEBUG nova.network.neutron [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Successfully updated port: fb56a4a1-033a-4205-aced-43f632d2f594 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 877.020247] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293826, 'name': CreateVM_Task, 'duration_secs': 0.525741} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.022930] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f0154308-e6a8-4321-a082-99291344664e] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 877.023889] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.024126] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.024435] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 877.024739] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4ad6efa-fcbf-45c6-a6e8-172c1de2af47 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.031565] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for the task: (returnval){ [ 877.031565] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]529a34be-57fe-2c4f-5748-1663724bc021" [ 877.031565] env[63021]: _type = "Task" [ 877.031565] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.043989] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]529a34be-57fe-2c4f-5748-1663724bc021, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.083536] env[63021]: DEBUG nova.compute.manager [req-a63f58b2-5b32-4997-924e-1b5e702273d5 req-07afce6c-3b94-4aa6-8d73-610ae625f8fb service nova] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Received event network-changed-fb56a4a1-033a-4205-aced-43f632d2f594 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 877.083536] env[63021]: DEBUG nova.compute.manager [req-a63f58b2-5b32-4997-924e-1b5e702273d5 req-07afce6c-3b94-4aa6-8d73-610ae625f8fb service nova] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Refreshing instance network info cache due to event network-changed-fb56a4a1-033a-4205-aced-43f632d2f594. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 877.083536] env[63021]: DEBUG oslo_concurrency.lockutils [req-a63f58b2-5b32-4997-924e-1b5e702273d5 req-07afce6c-3b94-4aa6-8d73-610ae625f8fb service nova] Acquiring lock "refresh_cache-44dadf8e-a727-448f-887d-6408f7ebfbc6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.083536] env[63021]: DEBUG oslo_concurrency.lockutils [req-a63f58b2-5b32-4997-924e-1b5e702273d5 req-07afce6c-3b94-4aa6-8d73-610ae625f8fb service nova] Acquired lock "refresh_cache-44dadf8e-a727-448f-887d-6408f7ebfbc6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.083536] env[63021]: DEBUG nova.network.neutron [req-a63f58b2-5b32-4997-924e-1b5e702273d5 req-07afce6c-3b94-4aa6-8d73-610ae625f8fb service nova] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Refreshing network info cache for port fb56a4a1-033a-4205-aced-43f632d2f594 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 877.131675] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-248b8d5a-a22d-4da5-a63d-2a6bf6999c2c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.141536] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f96cfddd-43ef-491e-aa20-81b39e9dd8db {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.174502] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf62bf54-e87a-44a4-9da9-58a2aee769e9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.184856] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bb1ca50-f113-4133-a42b-47ac2a75680d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.200227] env[63021]: DEBUG nova.compute.manager [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 877.200742] env[63021]: DEBUG nova.compute.provider_tree [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 877.203760] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae2cb16e-3240-4fa1-a18d-48a7c0eff7b5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.249161] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.343282] env[63021]: DEBUG nova.network.neutron [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Successfully updated port: 2f88c221-8f47-478e-a5f3-ba0a230c7ad4 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 877.370293] env[63021]: DEBUG nova.network.neutron [req-b2dc0cc9-a998-4967-909a-f7ff60fb5069 req-f68bd29d-0e1c-491e-a74a-e00cc06fa78d service nova] [instance: f0154308-e6a8-4321-a082-99291344664e] Updated VIF entry in instance network info cache for port ce58eb3e-a5f1-4d1a-ab2a-dd6fcb0d4f62. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 877.370703] env[63021]: DEBUG nova.network.neutron [req-b2dc0cc9-a998-4967-909a-f7ff60fb5069 req-f68bd29d-0e1c-491e-a74a-e00cc06fa78d service nova] [instance: f0154308-e6a8-4321-a082-99291344664e] Updating instance_info_cache with network_info: [{"id": "ce58eb3e-a5f1-4d1a-ab2a-dd6fcb0d4f62", "address": "fa:16:3e:12:e0:99", "network": {"id": "3e88ae40-5e21-4cbb-a0b4-84e0bd2910fa", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-557141593-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ed12e42e778445eb8f937906ce5ae72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce58eb3e-a5", "ovs_interfaceid": "ce58eb3e-a5f1-4d1a-ab2a-dd6fcb0d4f62", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.422426] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquiring lock "refresh_cache-44dadf8e-a727-448f-887d-6408f7ebfbc6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.542916] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]529a34be-57fe-2c4f-5748-1663724bc021, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.627432] env[63021]: DEBUG nova.network.neutron [req-a63f58b2-5b32-4997-924e-1b5e702273d5 req-07afce6c-3b94-4aa6-8d73-610ae625f8fb service nova] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 877.705187] env[63021]: DEBUG nova.scheduler.client.report [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 877.711764] env[63021]: DEBUG nova.network.neutron [req-a63f58b2-5b32-4997-924e-1b5e702273d5 req-07afce6c-3b94-4aa6-8d73-610ae625f8fb service nova] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.715878] env[63021]: INFO nova.compute.manager [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] instance snapshotting [ 877.718615] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2196cf3-1a14-4608-958e-4da42ef925d2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.748630] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cc86a2f-e158-42ba-a0bf-3d98e1ef5727 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.844727] env[63021]: DEBUG oslo_concurrency.lockutils [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "refresh_cache-36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.844937] env[63021]: DEBUG oslo_concurrency.lockutils [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquired lock "refresh_cache-36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.845138] env[63021]: DEBUG nova.network.neutron [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 877.874051] env[63021]: DEBUG oslo_concurrency.lockutils [req-b2dc0cc9-a998-4967-909a-f7ff60fb5069 req-f68bd29d-0e1c-491e-a74a-e00cc06fa78d service nova] Releasing lock "refresh_cache-f0154308-e6a8-4321-a082-99291344664e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.874389] env[63021]: DEBUG nova.compute.manager [req-b2dc0cc9-a998-4967-909a-f7ff60fb5069 req-f68bd29d-0e1c-491e-a74a-e00cc06fa78d service nova] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Received event network-vif-deleted-e581fd41-04b7-4081-a7be-c6389f4d623f {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 877.874623] env[63021]: DEBUG nova.compute.manager [req-b2dc0cc9-a998-4967-909a-f7ff60fb5069 req-f68bd29d-0e1c-491e-a74a-e00cc06fa78d service nova] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Received event network-vif-deleted-7f87a718-8fb1-40d8-8273-ebd48f298224 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 878.051487] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]529a34be-57fe-2c4f-5748-1663724bc021, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.212797] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.401s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.213445] env[63021]: DEBUG nova.compute.manager [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 878.216397] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 20.249s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.216616] env[63021]: DEBUG nova.objects.instance [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63021) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 878.221189] env[63021]: DEBUG oslo_concurrency.lockutils [req-a63f58b2-5b32-4997-924e-1b5e702273d5 req-07afce6c-3b94-4aa6-8d73-610ae625f8fb service nova] Releasing lock "refresh_cache-44dadf8e-a727-448f-887d-6408f7ebfbc6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.221899] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquired lock "refresh_cache-44dadf8e-a727-448f-887d-6408f7ebfbc6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.222095] env[63021]: DEBUG nova.network.neutron [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 878.269020] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Creating Snapshot of the VM instance {{(pid=63021) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 878.269020] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-cb71091a-d589-48de-b858-1d52dcffa65a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.280577] env[63021]: DEBUG oslo_vmware.api [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 878.280577] env[63021]: value = "task-1293827" [ 878.280577] env[63021]: _type = "Task" [ 878.280577] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.289772] env[63021]: DEBUG oslo_vmware.api [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293827, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.386212] env[63021]: WARNING nova.network.neutron [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] 2c4f9111-2f39-4bd0-95bc-09b96d535ea3 already exists in list: networks containing: ['2c4f9111-2f39-4bd0-95bc-09b96d535ea3']. ignoring it [ 878.386448] env[63021]: WARNING nova.network.neutron [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] 2c4f9111-2f39-4bd0-95bc-09b96d535ea3 already exists in list: networks containing: ['2c4f9111-2f39-4bd0-95bc-09b96d535ea3']. ignoring it [ 878.549825] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]529a34be-57fe-2c4f-5748-1663724bc021, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.724299] env[63021]: DEBUG nova.compute.utils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 878.735926] env[63021]: DEBUG nova.compute.manager [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 878.735926] env[63021]: DEBUG nova.network.neutron [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 878.794776] env[63021]: DEBUG oslo_vmware.api [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293827, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.809050] env[63021]: DEBUG nova.network.neutron [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 878.823530] env[63021]: DEBUG nova.compute.manager [req-63ad1206-3544-472c-93fd-38694952c997 req-4d16dead-2f9f-488b-970a-ec3ab2a786f6 service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Received event network-vif-plugged-2f88c221-8f47-478e-a5f3-ba0a230c7ad4 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 878.823828] env[63021]: DEBUG oslo_concurrency.lockutils [req-63ad1206-3544-472c-93fd-38694952c997 req-4d16dead-2f9f-488b-970a-ec3ab2a786f6 service nova] Acquiring lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.824018] env[63021]: DEBUG oslo_concurrency.lockutils [req-63ad1206-3544-472c-93fd-38694952c997 req-4d16dead-2f9f-488b-970a-ec3ab2a786f6 service nova] Lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.824240] env[63021]: DEBUG oslo_concurrency.lockutils [req-63ad1206-3544-472c-93fd-38694952c997 req-4d16dead-2f9f-488b-970a-ec3ab2a786f6 service nova] Lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.824427] env[63021]: DEBUG nova.compute.manager [req-63ad1206-3544-472c-93fd-38694952c997 req-4d16dead-2f9f-488b-970a-ec3ab2a786f6 service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] No waiting events found dispatching network-vif-plugged-2f88c221-8f47-478e-a5f3-ba0a230c7ad4 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 878.824669] env[63021]: WARNING nova.compute.manager [req-63ad1206-3544-472c-93fd-38694952c997 req-4d16dead-2f9f-488b-970a-ec3ab2a786f6 service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Received unexpected event network-vif-plugged-2f88c221-8f47-478e-a5f3-ba0a230c7ad4 for instance with vm_state active and task_state None. [ 878.824840] env[63021]: DEBUG nova.compute.manager [req-63ad1206-3544-472c-93fd-38694952c997 req-4d16dead-2f9f-488b-970a-ec3ab2a786f6 service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Received event network-changed-2f88c221-8f47-478e-a5f3-ba0a230c7ad4 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 878.824999] env[63021]: DEBUG nova.compute.manager [req-63ad1206-3544-472c-93fd-38694952c997 req-4d16dead-2f9f-488b-970a-ec3ab2a786f6 service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Refreshing instance network info cache due to event network-changed-2f88c221-8f47-478e-a5f3-ba0a230c7ad4. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 878.825203] env[63021]: DEBUG oslo_concurrency.lockutils [req-63ad1206-3544-472c-93fd-38694952c997 req-4d16dead-2f9f-488b-970a-ec3ab2a786f6 service nova] Acquiring lock "refresh_cache-36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.848089] env[63021]: DEBUG nova.policy [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '305f29bd44c949afa5731ec54c1a296d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1ed12e42e778445eb8f937906ce5ae72', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 878.911591] env[63021]: DEBUG nova.network.neutron [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Updating instance_info_cache with network_info: [{"id": "37dcb3f1-2931-40c6-96ba-b31b8fa1a72b", "address": "fa:16:3e:57:13:17", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37dcb3f1-29", "ovs_interfaceid": "37dcb3f1-2931-40c6-96ba-b31b8fa1a72b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d92edf1f-d15d-4f56-9da3-92cbe44b6709", "address": "fa:16:3e:b5:1b:c8", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd92edf1f-d1", "ovs_interfaceid": "d92edf1f-d15d-4f56-9da3-92cbe44b6709", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2f88c221-8f47-478e-a5f3-ba0a230c7ad4", "address": "fa:16:3e:c2:01:70", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f88c221-8f", "ovs_interfaceid": "2f88c221-8f47-478e-a5f3-ba0a230c7ad4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.052348] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]529a34be-57fe-2c4f-5748-1663724bc021, 'name': SearchDatastore_Task, 'duration_secs': 1.744151} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.052348] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.052348] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 879.052348] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.052348] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.052348] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 879.052585] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-56c36f49-3bad-4472-af66-6e57431f4da3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.065834] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 879.066044] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 879.066793] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-85404a4a-7fdc-409e-a0e4-b59c3188bf27 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.076591] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for the task: (returnval){ [ 879.076591] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5233d2b2-bf03-74da-da2c-d1b46867e4f0" [ 879.076591] env[63021]: _type = "Task" [ 879.076591] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.085615] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5233d2b2-bf03-74da-da2c-d1b46867e4f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.115213] env[63021]: DEBUG nova.network.neutron [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Updating instance_info_cache with network_info: [{"id": "fb56a4a1-033a-4205-aced-43f632d2f594", "address": "fa:16:3e:62:e9:e0", "network": {"id": "3e88ae40-5e21-4cbb-a0b4-84e0bd2910fa", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-557141593-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ed12e42e778445eb8f937906ce5ae72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb56a4a1-03", "ovs_interfaceid": "fb56a4a1-033a-4205-aced-43f632d2f594", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.232668] env[63021]: DEBUG nova.compute.manager [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 879.237965] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ca6ed7fd-3018-4193-8902-0796d4bfa2bc tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.022s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.239921] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.165s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.243709] env[63021]: INFO nova.compute.claims [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 879.253028] env[63021]: DEBUG nova.network.neutron [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Successfully created port: 71541497-0e60-4ea5-9d4a-33970a716736 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 879.291258] env[63021]: DEBUG oslo_vmware.api [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293827, 'name': CreateSnapshot_Task, 'duration_secs': 0.783524} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.291659] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Created Snapshot of the VM instance {{(pid=63021) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 879.292402] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-869c9c57-ff37-4c79-b9da-6dbb2f77dacd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.415489] env[63021]: DEBUG oslo_concurrency.lockutils [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Releasing lock "refresh_cache-36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.416575] env[63021]: DEBUG oslo_concurrency.lockutils [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.416854] env[63021]: DEBUG oslo_concurrency.lockutils [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquired lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.417247] env[63021]: DEBUG oslo_concurrency.lockutils [req-63ad1206-3544-472c-93fd-38694952c997 req-4d16dead-2f9f-488b-970a-ec3ab2a786f6 service nova] Acquired lock "refresh_cache-36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.417438] env[63021]: DEBUG nova.network.neutron [req-63ad1206-3544-472c-93fd-38694952c997 req-4d16dead-2f9f-488b-970a-ec3ab2a786f6 service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Refreshing network info cache for port 2f88c221-8f47-478e-a5f3-ba0a230c7ad4 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 879.419161] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03628a16-04ad-4225-a140-56ffcba999c7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.442447] env[63021]: DEBUG nova.virt.hardware [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 879.442679] env[63021]: DEBUG nova.virt.hardware [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 879.442919] env[63021]: DEBUG nova.virt.hardware [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 879.443010] env[63021]: DEBUG nova.virt.hardware [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 879.443183] env[63021]: DEBUG nova.virt.hardware [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 879.443328] env[63021]: DEBUG nova.virt.hardware [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 879.443526] env[63021]: DEBUG nova.virt.hardware [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 879.443677] env[63021]: DEBUG nova.virt.hardware [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 879.443836] env[63021]: DEBUG nova.virt.hardware [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 879.443995] env[63021]: DEBUG nova.virt.hardware [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 879.444642] env[63021]: DEBUG nova.virt.hardware [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 879.451183] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Reconfiguring VM to attach interface {{(pid=63021) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 879.451793] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6f3ce967-c61a-48fd-9cbe-08a18cc7cfcc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.470593] env[63021]: DEBUG oslo_vmware.api [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 879.470593] env[63021]: value = "task-1293828" [ 879.470593] env[63021]: _type = "Task" [ 879.470593] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.479019] env[63021]: DEBUG oslo_vmware.api [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293828, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.589839] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5233d2b2-bf03-74da-da2c-d1b46867e4f0, 'name': SearchDatastore_Task, 'duration_secs': 0.011297} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.590657] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa8178da-1c78-4cc8-91e6-7a95fcfbb47d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.597945] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for the task: (returnval){ [ 879.597945] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52703805-7282-c71b-abb7-e37024327f96" [ 879.597945] env[63021]: _type = "Task" [ 879.597945] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.606970] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52703805-7282-c71b-abb7-e37024327f96, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.620159] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Releasing lock "refresh_cache-44dadf8e-a727-448f-887d-6408f7ebfbc6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.620159] env[63021]: DEBUG nova.compute.manager [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Instance network_info: |[{"id": "fb56a4a1-033a-4205-aced-43f632d2f594", "address": "fa:16:3e:62:e9:e0", "network": {"id": "3e88ae40-5e21-4cbb-a0b4-84e0bd2910fa", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-557141593-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ed12e42e778445eb8f937906ce5ae72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb56a4a1-03", "ovs_interfaceid": "fb56a4a1-033a-4205-aced-43f632d2f594", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 879.620159] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:62:e9:e0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b1f3e6c3-5584-4852-9017-476ab8ac4946', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fb56a4a1-033a-4205-aced-43f632d2f594', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 879.626177] env[63021]: DEBUG oslo.service.loopingcall [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 879.626539] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 879.626875] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7f660a9c-6cd6-4120-b1e7-e39f785fc3a3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.650043] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 879.650043] env[63021]: value = "task-1293829" [ 879.650043] env[63021]: _type = "Task" [ 879.650043] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.659666] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293829, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.814392] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Creating linked-clone VM from snapshot {{(pid=63021) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 879.816463] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-8dc5c9ca-e4a8-4e42-b262-e20d4afc0893 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.825443] env[63021]: DEBUG oslo_vmware.api [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 879.825443] env[63021]: value = "task-1293830" [ 879.825443] env[63021]: _type = "Task" [ 879.825443] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.835800] env[63021]: DEBUG oslo_vmware.api [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293830, 'name': CloneVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.984035] env[63021]: DEBUG oslo_vmware.api [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293828, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.114053] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52703805-7282-c71b-abb7-e37024327f96, 'name': SearchDatastore_Task, 'duration_secs': 0.010659} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.114459] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.114746] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] f0154308-e6a8-4321-a082-99291344664e/f0154308-e6a8-4321-a082-99291344664e.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 880.115084] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5ffa96b4-bde6-49da-be5e-0944a60ec74c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.125280] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for the task: (returnval){ [ 880.125280] env[63021]: value = "task-1293831" [ 880.125280] env[63021]: _type = "Task" [ 880.125280] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.137796] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293831, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.161064] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293829, 'name': CreateVM_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.241963] env[63021]: DEBUG nova.network.neutron [req-63ad1206-3544-472c-93fd-38694952c997 req-4d16dead-2f9f-488b-970a-ec3ab2a786f6 service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Updated VIF entry in instance network info cache for port 2f88c221-8f47-478e-a5f3-ba0a230c7ad4. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 880.242445] env[63021]: DEBUG nova.network.neutron [req-63ad1206-3544-472c-93fd-38694952c997 req-4d16dead-2f9f-488b-970a-ec3ab2a786f6 service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Updating instance_info_cache with network_info: [{"id": "37dcb3f1-2931-40c6-96ba-b31b8fa1a72b", "address": "fa:16:3e:57:13:17", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37dcb3f1-29", "ovs_interfaceid": "37dcb3f1-2931-40c6-96ba-b31b8fa1a72b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d92edf1f-d15d-4f56-9da3-92cbe44b6709", "address": "fa:16:3e:b5:1b:c8", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd92edf1f-d1", "ovs_interfaceid": "d92edf1f-d15d-4f56-9da3-92cbe44b6709", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2f88c221-8f47-478e-a5f3-ba0a230c7ad4", "address": "fa:16:3e:c2:01:70", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f88c221-8f", "ovs_interfaceid": "2f88c221-8f47-478e-a5f3-ba0a230c7ad4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.245035] env[63021]: DEBUG nova.compute.manager [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 880.270630] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 880.270878] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 880.271042] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 880.271224] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 880.271372] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 880.271519] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 880.271719] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 880.271881] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 880.272074] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 880.272252] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 880.272456] env[63021]: DEBUG nova.virt.hardware [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 880.273694] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-048df14e-8b2f-4bf7-8896-01a424a6ce38 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.282710] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e6fe40b-b098-4998-a18a-3079d9542743 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.338728] env[63021]: DEBUG oslo_vmware.api [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293830, 'name': CloneVM_Task} progress is 94%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.488620] env[63021]: DEBUG oslo_vmware.api [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293828, 'name': ReconfigVM_Task, 'duration_secs': 0.640948} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.493128] env[63021]: DEBUG oslo_concurrency.lockutils [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Releasing lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.493387] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Reconfigured VM to attach interface {{(pid=63021) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 880.581640] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70b06845-3e87-4b83-8f0d-fde5e06683f2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.591523] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e18fe7f8-1b26-42a9-9787-b58c226308bc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.630942] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a01c01ba-b84b-47ee-80b7-09b3d80db14e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.640299] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293831, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.643777] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e16488b-5d9a-413f-b92a-f632eae27dfd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.659356] env[63021]: DEBUG nova.compute.provider_tree [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 880.671011] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293829, 'name': CreateVM_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.749036] env[63021]: DEBUG oslo_concurrency.lockutils [req-63ad1206-3544-472c-93fd-38694952c997 req-4d16dead-2f9f-488b-970a-ec3ab2a786f6 service nova] Releasing lock "refresh_cache-36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.777382] env[63021]: DEBUG nova.network.neutron [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Successfully updated port: 71541497-0e60-4ea5-9d4a-33970a716736 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 880.838907] env[63021]: DEBUG oslo_vmware.api [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293830, 'name': CloneVM_Task} progress is 94%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.905976] env[63021]: DEBUG nova.compute.manager [req-ccb5dc1a-4dcb-4a17-b176-15b81748e399 req-366e5f97-9952-4559-838e-fa57afcad998 service nova] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Received event network-vif-plugged-71541497-0e60-4ea5-9d4a-33970a716736 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 880.906203] env[63021]: DEBUG oslo_concurrency.lockutils [req-ccb5dc1a-4dcb-4a17-b176-15b81748e399 req-366e5f97-9952-4559-838e-fa57afcad998 service nova] Acquiring lock "50a1269e-4d83-4cf7-ba14-3b1afb3c43ed-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.906448] env[63021]: DEBUG oslo_concurrency.lockutils [req-ccb5dc1a-4dcb-4a17-b176-15b81748e399 req-366e5f97-9952-4559-838e-fa57afcad998 service nova] Lock "50a1269e-4d83-4cf7-ba14-3b1afb3c43ed-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.906563] env[63021]: DEBUG oslo_concurrency.lockutils [req-ccb5dc1a-4dcb-4a17-b176-15b81748e399 req-366e5f97-9952-4559-838e-fa57afcad998 service nova] Lock "50a1269e-4d83-4cf7-ba14-3b1afb3c43ed-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.906727] env[63021]: DEBUG nova.compute.manager [req-ccb5dc1a-4dcb-4a17-b176-15b81748e399 req-366e5f97-9952-4559-838e-fa57afcad998 service nova] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] No waiting events found dispatching network-vif-plugged-71541497-0e60-4ea5-9d4a-33970a716736 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 880.906941] env[63021]: WARNING nova.compute.manager [req-ccb5dc1a-4dcb-4a17-b176-15b81748e399 req-366e5f97-9952-4559-838e-fa57afcad998 service nova] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Received unexpected event network-vif-plugged-71541497-0e60-4ea5-9d4a-33970a716736 for instance with vm_state building and task_state spawning. [ 880.907134] env[63021]: DEBUG nova.compute.manager [req-ccb5dc1a-4dcb-4a17-b176-15b81748e399 req-366e5f97-9952-4559-838e-fa57afcad998 service nova] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Received event network-changed-71541497-0e60-4ea5-9d4a-33970a716736 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 880.907288] env[63021]: DEBUG nova.compute.manager [req-ccb5dc1a-4dcb-4a17-b176-15b81748e399 req-366e5f97-9952-4559-838e-fa57afcad998 service nova] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Refreshing instance network info cache due to event network-changed-71541497-0e60-4ea5-9d4a-33970a716736. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 880.907467] env[63021]: DEBUG oslo_concurrency.lockutils [req-ccb5dc1a-4dcb-4a17-b176-15b81748e399 req-366e5f97-9952-4559-838e-fa57afcad998 service nova] Acquiring lock "refresh_cache-50a1269e-4d83-4cf7-ba14-3b1afb3c43ed" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.907597] env[63021]: DEBUG oslo_concurrency.lockutils [req-ccb5dc1a-4dcb-4a17-b176-15b81748e399 req-366e5f97-9952-4559-838e-fa57afcad998 service nova] Acquired lock "refresh_cache-50a1269e-4d83-4cf7-ba14-3b1afb3c43ed" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.907751] env[63021]: DEBUG nova.network.neutron [req-ccb5dc1a-4dcb-4a17-b176-15b81748e399 req-366e5f97-9952-4559-838e-fa57afcad998 service nova] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Refreshing network info cache for port 71541497-0e60-4ea5-9d4a-33970a716736 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 881.001488] env[63021]: DEBUG oslo_concurrency.lockutils [None req-97e10d18-69a5-4cf8-9e59-17a11048e1e5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "interface-36ec45ec-761b-4d62-a74f-e4d9a840ada0-2f88c221-8f47-478e-a5f3-ba0a230c7ad4" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.815s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.138215] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293831, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.581193} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.138522] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] f0154308-e6a8-4321-a082-99291344664e/f0154308-e6a8-4321-a082-99291344664e.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 881.138751] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 881.139037] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1de269f8-09fe-4f7d-b782-7a657c0463e1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.146691] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for the task: (returnval){ [ 881.146691] env[63021]: value = "task-1293832" [ 881.146691] env[63021]: _type = "Task" [ 881.146691] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.155647] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293832, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.165251] env[63021]: DEBUG nova.scheduler.client.report [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 881.172063] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293829, 'name': CreateVM_Task, 'duration_secs': 1.473866} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.172441] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 881.173163] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.173409] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.173699] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 881.173974] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bebd761c-2a13-468b-a462-24ad0995c30a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.179796] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for the task: (returnval){ [ 881.179796] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5298c15f-4f69-0369-7acd-1e4bd7d098f9" [ 881.179796] env[63021]: _type = "Task" [ 881.179796] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.191133] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5298c15f-4f69-0369-7acd-1e4bd7d098f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.282970] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquiring lock "refresh_cache-50a1269e-4d83-4cf7-ba14-3b1afb3c43ed" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.339711] env[63021]: DEBUG oslo_vmware.api [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293830, 'name': CloneVM_Task} progress is 95%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.440454] env[63021]: DEBUG nova.network.neutron [req-ccb5dc1a-4dcb-4a17-b176-15b81748e399 req-366e5f97-9952-4559-838e-fa57afcad998 service nova] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 881.658785] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293832, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.128727} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.659084] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 881.659857] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-198fb1b1-4a75-4fc8-a2dd-cd53e5929ad6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.676623] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.437s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.677096] env[63021]: DEBUG nova.compute.manager [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 881.687798] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] f0154308-e6a8-4321-a082-99291344664e/f0154308-e6a8-4321-a082-99291344664e.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 881.688686] env[63021]: DEBUG nova.network.neutron [req-ccb5dc1a-4dcb-4a17-b176-15b81748e399 req-366e5f97-9952-4559-838e-fa57afcad998 service nova] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.690151] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.603s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.691328] env[63021]: INFO nova.compute.claims [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 881.695312] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9eccf3cc-6d29-41ee-9caa-ec1ffd5e3d65 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.720366] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5298c15f-4f69-0369-7acd-1e4bd7d098f9, 'name': SearchDatastore_Task, 'duration_secs': 0.011109} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.721676] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.721915] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 881.722162] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.722305] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.722486] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 881.723093] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for the task: (returnval){ [ 881.723093] env[63021]: value = "task-1293833" [ 881.723093] env[63021]: _type = "Task" [ 881.723093] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.723289] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ce42d06c-de4b-4cae-8a7f-156eb49c3a66 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.734877] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293833, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.736324] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 881.736531] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 881.737333] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2701e7e1-9c9b-49c2-a647-a70849edf45c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.743181] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for the task: (returnval){ [ 881.743181] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]528116a9-9413-60f0-979d-25e36caa8c09" [ 881.743181] env[63021]: _type = "Task" [ 881.743181] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.752558] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]528116a9-9413-60f0-979d-25e36caa8c09, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.841520] env[63021]: DEBUG oslo_vmware.api [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293830, 'name': CloneVM_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.192675] env[63021]: DEBUG nova.compute.utils [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 882.193847] env[63021]: DEBUG nova.compute.manager [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 882.194033] env[63021]: DEBUG nova.network.neutron [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 882.211785] env[63021]: DEBUG oslo_concurrency.lockutils [req-ccb5dc1a-4dcb-4a17-b176-15b81748e399 req-366e5f97-9952-4559-838e-fa57afcad998 service nova] Releasing lock "refresh_cache-50a1269e-4d83-4cf7-ba14-3b1afb3c43ed" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.212510] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquired lock "refresh_cache-50a1269e-4d83-4cf7-ba14-3b1afb3c43ed" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.212698] env[63021]: DEBUG nova.network.neutron [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 882.236405] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293833, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.254807] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]528116a9-9413-60f0-979d-25e36caa8c09, 'name': SearchDatastore_Task, 'duration_secs': 0.010739} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.256044] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3587a95e-334c-4c0e-ba7b-93d1b66607a3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.262335] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for the task: (returnval){ [ 882.262335] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52cba609-b645-9af3-9bbb-8512591ee632" [ 882.262335] env[63021]: _type = "Task" [ 882.262335] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.266842] env[63021]: DEBUG nova.policy [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '14d6865c04e44c069a31ca014c2a332d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3f3d9b0b38e647a6b13f2a9af471147b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 882.274274] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52cba609-b645-9af3-9bbb-8512591ee632, 'name': SearchDatastore_Task, 'duration_secs': 0.009445} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.274521] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.274905] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 44dadf8e-a727-448f-887d-6408f7ebfbc6/44dadf8e-a727-448f-887d-6408f7ebfbc6.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 882.275136] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2f7a3ff7-b63d-4b7b-8dfa-b419338b3b29 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.281731] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for the task: (returnval){ [ 882.281731] env[63021]: value = "task-1293834" [ 882.281731] env[63021]: _type = "Task" [ 882.281731] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.290142] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293834, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.342598] env[63021]: DEBUG oslo_vmware.api [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293830, 'name': CloneVM_Task, 'duration_secs': 2.139978} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.342949] env[63021]: INFO nova.virt.vmwareapi.vmops [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Created linked-clone VM from snapshot [ 882.343791] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2916559a-d090-48a5-a67e-113bbdf98dc5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.352820] env[63021]: DEBUG nova.virt.vmwareapi.images [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Uploading image 02de5fc5-60f6-49e2-ae03-e14c840dae1b {{(pid=63021) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 882.379843] env[63021]: DEBUG oslo_vmware.rw_handles [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 882.379843] env[63021]: value = "vm-277530" [ 882.379843] env[63021]: _type = "VirtualMachine" [ 882.379843] env[63021]: }. {{(pid=63021) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 882.380207] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-77343be6-7f30-4100-b316-4e34eefbab1f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.388602] env[63021]: DEBUG oslo_vmware.rw_handles [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lease: (returnval){ [ 882.388602] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52a82cbe-68f7-829f-7b8a-1894bcd88160" [ 882.388602] env[63021]: _type = "HttpNfcLease" [ 882.388602] env[63021]: } obtained for exporting VM: (result){ [ 882.388602] env[63021]: value = "vm-277530" [ 882.388602] env[63021]: _type = "VirtualMachine" [ 882.388602] env[63021]: }. {{(pid=63021) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 882.389015] env[63021]: DEBUG oslo_vmware.api [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the lease: (returnval){ [ 882.389015] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52a82cbe-68f7-829f-7b8a-1894bcd88160" [ 882.389015] env[63021]: _type = "HttpNfcLease" [ 882.389015] env[63021]: } to be ready. {{(pid=63021) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 882.396990] env[63021]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 882.396990] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52a82cbe-68f7-829f-7b8a-1894bcd88160" [ 882.396990] env[63021]: _type = "HttpNfcLease" [ 882.396990] env[63021]: } is initializing. {{(pid=63021) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 882.555200] env[63021]: DEBUG nova.network.neutron [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Successfully created port: 664021c6-82fb-46ea-a7d2-c9e942e0dded {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 882.587950] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dcf93f26-6ad3-471b-b85c-2f39720207f0 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "interface-36ec45ec-761b-4d62-a74f-e4d9a840ada0-d92edf1f-d15d-4f56-9da3-92cbe44b6709" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.588251] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dcf93f26-6ad3-471b-b85c-2f39720207f0 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "interface-36ec45ec-761b-4d62-a74f-e4d9a840ada0-d92edf1f-d15d-4f56-9da3-92cbe44b6709" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.697500] env[63021]: DEBUG nova.compute.manager [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 882.739944] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293833, 'name': ReconfigVM_Task, 'duration_secs': 0.773729} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.743283] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Reconfigured VM instance instance-00000048 to attach disk [datastore1] f0154308-e6a8-4321-a082-99291344664e/f0154308-e6a8-4321-a082-99291344664e.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 882.745479] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b60108ed-6b76-4a33-a490-0a9af5698a6a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.754814] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for the task: (returnval){ [ 882.754814] env[63021]: value = "task-1293836" [ 882.754814] env[63021]: _type = "Task" [ 882.754814] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.774495] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293836, 'name': Rename_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.775807] env[63021]: DEBUG nova.network.neutron [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 882.794705] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293834, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.902383] env[63021]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 882.902383] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52a82cbe-68f7-829f-7b8a-1894bcd88160" [ 882.902383] env[63021]: _type = "HttpNfcLease" [ 882.902383] env[63021]: } is ready. {{(pid=63021) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 882.905017] env[63021]: DEBUG oslo_vmware.rw_handles [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 882.905017] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52a82cbe-68f7-829f-7b8a-1894bcd88160" [ 882.905017] env[63021]: _type = "HttpNfcLease" [ 882.905017] env[63021]: }. {{(pid=63021) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 882.906017] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c578cddc-2023-49d3-90e7-da452bb22575 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.914882] env[63021]: DEBUG oslo_vmware.rw_handles [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5230102e-7773-2188-3adf-9cbf4ac916bf/disk-0.vmdk from lease info. {{(pid=63021) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 882.915078] env[63021]: DEBUG oslo_vmware.rw_handles [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5230102e-7773-2188-3adf-9cbf4ac916bf/disk-0.vmdk for reading. {{(pid=63021) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 883.014954] env[63021]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-89602aa7-1a46-4bd8-89fb-95835ac71ffa {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.021468] env[63021]: DEBUG nova.network.neutron [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Updating instance_info_cache with network_info: [{"id": "71541497-0e60-4ea5-9d4a-33970a716736", "address": "fa:16:3e:25:4b:22", "network": {"id": "3e88ae40-5e21-4cbb-a0b4-84e0bd2910fa", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-557141593-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ed12e42e778445eb8f937906ce5ae72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap71541497-0e", "ovs_interfaceid": "71541497-0e60-4ea5-9d4a-33970a716736", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.089685] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cf16faa-2f10-468c-9a17-32c04d6c217e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.094137] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dcf93f26-6ad3-471b-b85c-2f39720207f0 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.094137] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dcf93f26-6ad3-471b-b85c-2f39720207f0 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquired lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.094137] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62f2c27f-c6bf-403e-94be-84cd3356d346 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.114117] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09de6720-20d8-4654-8a45-b1ba7e13f3bc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.118509] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c48a544-d6e2-46cd-8df3-ac334f58c4f1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.177861] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-dcf93f26-6ad3-471b-b85c-2f39720207f0 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Reconfiguring VM to detach interface {{(pid=63021) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 883.178573] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0f66495d-c509-4c31-ad8c-08597ac2046d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.195786] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1a9a5f3-6607-4f72-8a3d-71c4068559f1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.202945] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18647541-59ad-4d04-8dcd-57a7574de3b0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.210408] env[63021]: DEBUG oslo_vmware.api [None req-dcf93f26-6ad3-471b-b85c-2f39720207f0 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 883.210408] env[63021]: value = "task-1293837" [ 883.210408] env[63021]: _type = "Task" [ 883.210408] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.223244] env[63021]: DEBUG nova.compute.provider_tree [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 883.233695] env[63021]: DEBUG oslo_vmware.api [None req-dcf93f26-6ad3-471b-b85c-2f39720207f0 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293837, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.266035] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293836, 'name': Rename_Task, 'duration_secs': 0.181273} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.266614] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 883.267550] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-020c1023-528c-45be-a232-5e21fae35067 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.275996] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for the task: (returnval){ [ 883.275996] env[63021]: value = "task-1293838" [ 883.275996] env[63021]: _type = "Task" [ 883.275996] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.284885] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293838, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.293959] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293834, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.585021} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.294262] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 44dadf8e-a727-448f-887d-6408f7ebfbc6/44dadf8e-a727-448f-887d-6408f7ebfbc6.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 883.294485] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 883.294776] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-032e177e-2057-4ba5-8ee7-186d0a0f7bc7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.303126] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for the task: (returnval){ [ 883.303126] env[63021]: value = "task-1293839" [ 883.303126] env[63021]: _type = "Task" [ 883.303126] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.312921] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293839, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.525737] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Releasing lock "refresh_cache-50a1269e-4d83-4cf7-ba14-3b1afb3c43ed" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.525990] env[63021]: DEBUG nova.compute.manager [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Instance network_info: |[{"id": "71541497-0e60-4ea5-9d4a-33970a716736", "address": "fa:16:3e:25:4b:22", "network": {"id": "3e88ae40-5e21-4cbb-a0b4-84e0bd2910fa", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-557141593-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ed12e42e778445eb8f937906ce5ae72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap71541497-0e", "ovs_interfaceid": "71541497-0e60-4ea5-9d4a-33970a716736", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 883.527031] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:25:4b:22', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b1f3e6c3-5584-4852-9017-476ab8ac4946', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '71541497-0e60-4ea5-9d4a-33970a716736', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 883.539411] env[63021]: DEBUG oslo.service.loopingcall [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 883.539813] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 883.540186] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2c75b63f-65b1-4189-a81a-781375e0e0ec {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.572027] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 883.572027] env[63021]: value = "task-1293840" [ 883.572027] env[63021]: _type = "Task" [ 883.572027] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.582857] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293840, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.712472] env[63021]: DEBUG nova.compute.manager [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 883.725552] env[63021]: DEBUG oslo_vmware.api [None req-dcf93f26-6ad3-471b-b85c-2f39720207f0 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293837, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.730658] env[63021]: DEBUG nova.scheduler.client.report [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 883.741146] env[63021]: DEBUG nova.virt.hardware [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 883.741410] env[63021]: DEBUG nova.virt.hardware [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 883.741577] env[63021]: DEBUG nova.virt.hardware [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 883.741774] env[63021]: DEBUG nova.virt.hardware [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 883.742359] env[63021]: DEBUG nova.virt.hardware [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 883.742549] env[63021]: DEBUG nova.virt.hardware [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 883.742773] env[63021]: DEBUG nova.virt.hardware [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 883.742941] env[63021]: DEBUG nova.virt.hardware [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 883.743134] env[63021]: DEBUG nova.virt.hardware [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 883.743333] env[63021]: DEBUG nova.virt.hardware [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 883.743572] env[63021]: DEBUG nova.virt.hardware [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 883.744565] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afc038aa-b652-4a20-b98f-1a1607543272 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.754420] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbeafb5b-c7eb-48d0-8ea1-c35ac628b819 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.787264] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293838, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.814955] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293839, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089643} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.815399] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 883.816341] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78a36118-b99d-49d6-a53d-f4a5b2e437fe {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.843888] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] 44dadf8e-a727-448f-887d-6408f7ebfbc6/44dadf8e-a727-448f-887d-6408f7ebfbc6.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 883.845780] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-be440045-ccf2-40b5-933f-662b2762bf7e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.867560] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for the task: (returnval){ [ 883.867560] env[63021]: value = "task-1293841" [ 883.867560] env[63021]: _type = "Task" [ 883.867560] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.877991] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293841, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.082650] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293840, 'name': CreateVM_Task, 'duration_secs': 0.44728} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.083168] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 884.084011] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.084286] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.084661] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 884.084985] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed3a57df-83e5-49eb-94fd-93ebbc4d967d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.091205] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for the task: (returnval){ [ 884.091205] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]522f1bd8-643d-ebf8-e65c-9a3d1b9b3b63" [ 884.091205] env[63021]: _type = "Task" [ 884.091205] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.100531] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]522f1bd8-643d-ebf8-e65c-9a3d1b9b3b63, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.111548] env[63021]: DEBUG nova.compute.manager [req-e8bb9b7d-9820-423a-9398-e4b856b4517c req-ea6a85df-9031-4450-94e1-1fc161e88efe service nova] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Received event network-vif-plugged-664021c6-82fb-46ea-a7d2-c9e942e0dded {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 884.111963] env[63021]: DEBUG oslo_concurrency.lockutils [req-e8bb9b7d-9820-423a-9398-e4b856b4517c req-ea6a85df-9031-4450-94e1-1fc161e88efe service nova] Acquiring lock "51e4a14c-4792-43cd-9e99-4e06a8f6d6fb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.112223] env[63021]: DEBUG oslo_concurrency.lockutils [req-e8bb9b7d-9820-423a-9398-e4b856b4517c req-ea6a85df-9031-4450-94e1-1fc161e88efe service nova] Lock "51e4a14c-4792-43cd-9e99-4e06a8f6d6fb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.112460] env[63021]: DEBUG oslo_concurrency.lockutils [req-e8bb9b7d-9820-423a-9398-e4b856b4517c req-ea6a85df-9031-4450-94e1-1fc161e88efe service nova] Lock "51e4a14c-4792-43cd-9e99-4e06a8f6d6fb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.112776] env[63021]: DEBUG nova.compute.manager [req-e8bb9b7d-9820-423a-9398-e4b856b4517c req-ea6a85df-9031-4450-94e1-1fc161e88efe service nova] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] No waiting events found dispatching network-vif-plugged-664021c6-82fb-46ea-a7d2-c9e942e0dded {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 884.113008] env[63021]: WARNING nova.compute.manager [req-e8bb9b7d-9820-423a-9398-e4b856b4517c req-ea6a85df-9031-4450-94e1-1fc161e88efe service nova] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Received unexpected event network-vif-plugged-664021c6-82fb-46ea-a7d2-c9e942e0dded for instance with vm_state building and task_state spawning. [ 884.215610] env[63021]: DEBUG nova.network.neutron [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Successfully updated port: 664021c6-82fb-46ea-a7d2-c9e942e0dded {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 884.227797] env[63021]: DEBUG oslo_vmware.api [None req-dcf93f26-6ad3-471b-b85c-2f39720207f0 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293837, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.236184] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.546s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.236712] env[63021]: DEBUG nova.compute.manager [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 884.239337] env[63021]: DEBUG oslo_concurrency.lockutils [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.477s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.241187] env[63021]: INFO nova.compute.claims [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 884.290978] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293838, 'name': PowerOnVM_Task, 'duration_secs': 0.792062} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.291288] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 884.291642] env[63021]: INFO nova.compute.manager [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Took 10.86 seconds to spawn the instance on the hypervisor. [ 884.291909] env[63021]: DEBUG nova.compute.manager [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 884.292942] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e1f27e3-4152-4c66-a5ad-9275fa654844 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.377935] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293841, 'name': ReconfigVM_Task, 'duration_secs': 0.320401} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.378258] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Reconfigured VM instance instance-00000049 to attach disk [datastore1] 44dadf8e-a727-448f-887d-6408f7ebfbc6/44dadf8e-a727-448f-887d-6408f7ebfbc6.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 884.378908] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-482621b5-780b-4d7c-8dd5-5eb5270f3c1b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.386535] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for the task: (returnval){ [ 884.386535] env[63021]: value = "task-1293842" [ 884.386535] env[63021]: _type = "Task" [ 884.386535] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.396209] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293842, 'name': Rename_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.603538] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]522f1bd8-643d-ebf8-e65c-9a3d1b9b3b63, 'name': SearchDatastore_Task, 'duration_secs': 0.011517} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.603979] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.604293] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 884.604590] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.605022] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.605300] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 884.605660] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-96641c94-56ae-49d6-bdc8-9560dc4bdbef {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.615593] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 884.617012] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 884.617012] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17ed39e2-50b7-4edf-98e1-89bb508050b3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.624186] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for the task: (returnval){ [ 884.624186] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52b5e64a-1fe3-4579-6ee7-06e35fa04815" [ 884.624186] env[63021]: _type = "Task" [ 884.624186] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.634337] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52b5e64a-1fe3-4579-6ee7-06e35fa04815, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.723871] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Acquiring lock "refresh_cache-51e4a14c-4792-43cd-9e99-4e06a8f6d6fb" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.724129] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Acquired lock "refresh_cache-51e4a14c-4792-43cd-9e99-4e06a8f6d6fb" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.724351] env[63021]: DEBUG nova.network.neutron [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 884.734408] env[63021]: DEBUG oslo_vmware.api [None req-dcf93f26-6ad3-471b-b85c-2f39720207f0 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293837, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.749050] env[63021]: DEBUG nova.compute.utils [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 884.751140] env[63021]: DEBUG nova.compute.manager [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 884.751268] env[63021]: DEBUG nova.network.neutron [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 884.796839] env[63021]: DEBUG nova.policy [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4fe657b588214bc29667fe9453a61038', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2d364c9eae4b4f8a8b60d4a3ac2ed648', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 884.815172] env[63021]: INFO nova.compute.manager [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Took 39.13 seconds to build instance. [ 884.901399] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293842, 'name': Rename_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.068749] env[63021]: DEBUG nova.network.neutron [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Successfully created port: d34cfd0d-464e-400b-8366-83321cc6112f {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 885.136064] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52b5e64a-1fe3-4579-6ee7-06e35fa04815, 'name': SearchDatastore_Task, 'duration_secs': 0.010444} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.136911] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6313d637-755f-4f20-a7b1-b700ac839cc1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.143991] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for the task: (returnval){ [ 885.143991] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52c4f008-80c1-5caf-ac86-29e61ce0f2eb" [ 885.143991] env[63021]: _type = "Task" [ 885.143991] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.153786] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52c4f008-80c1-5caf-ac86-29e61ce0f2eb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.230519] env[63021]: DEBUG oslo_vmware.api [None req-dcf93f26-6ad3-471b-b85c-2f39720207f0 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293837, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.254900] env[63021]: DEBUG nova.compute.manager [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 885.277268] env[63021]: DEBUG nova.network.neutron [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 885.316676] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Lock "f0154308-e6a8-4321-a082-99291344664e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 79.895s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.401184] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293842, 'name': Rename_Task, 'duration_secs': 0.896172} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.403860] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 885.404347] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1a278522-4da3-4194-93bc-0306e782fce2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.411901] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for the task: (returnval){ [ 885.411901] env[63021]: value = "task-1293843" [ 885.411901] env[63021]: _type = "Task" [ 885.411901] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.422710] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293843, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.466122] env[63021]: DEBUG nova.network.neutron [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Updating instance_info_cache with network_info: [{"id": "664021c6-82fb-46ea-a7d2-c9e942e0dded", "address": "fa:16:3e:3d:ca:1e", "network": {"id": "d67a9f2c-02da-411f-81dd-e2d702d6dd8b", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.84", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "cc50c08b679248798f76b07c1530ed7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "503991c4-44d0-42d9-aa03-5259331f1051", "external-id": "nsx-vlan-transportzone-3", "segmentation_id": 3, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap664021c6-82", "ovs_interfaceid": "664021c6-82fb-46ea-a7d2-c9e942e0dded", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.560466] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e13a28a-bab4-45a8-add1-92666c6eb221 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.569592] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb860c87-7475-4299-b5d7-f686dfb17ee6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.602290] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18df3da7-2783-4bfa-b191-7bc0260452a0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.610918] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81d5f982-7020-4a77-b5db-1c506c98509d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.625890] env[63021]: DEBUG nova.compute.provider_tree [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 885.657026] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52c4f008-80c1-5caf-ac86-29e61ce0f2eb, 'name': SearchDatastore_Task, 'duration_secs': 0.012534} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.657026] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.657026] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed/50a1269e-4d83-4cf7-ba14-3b1afb3c43ed.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 885.657026] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-99d42e14-44c8-4470-8db5-abf63cfcc3e7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.666019] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for the task: (returnval){ [ 885.666019] env[63021]: value = "task-1293844" [ 885.666019] env[63021]: _type = "Task" [ 885.666019] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.675858] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293844, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.728868] env[63021]: DEBUG oslo_vmware.api [None req-dcf93f26-6ad3-471b-b85c-2f39720207f0 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293837, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.922641] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293843, 'name': PowerOnVM_Task, 'duration_secs': 0.470838} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.923094] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 885.923421] env[63021]: INFO nova.compute.manager [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Took 10.01 seconds to spawn the instance on the hypervisor. [ 885.923740] env[63021]: DEBUG nova.compute.manager [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 885.924781] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e79edb3b-36af-45df-96b5-229c0f78ce89 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.969364] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Releasing lock "refresh_cache-51e4a14c-4792-43cd-9e99-4e06a8f6d6fb" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.969951] env[63021]: DEBUG nova.compute.manager [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Instance network_info: |[{"id": "664021c6-82fb-46ea-a7d2-c9e942e0dded", "address": "fa:16:3e:3d:ca:1e", "network": {"id": "d67a9f2c-02da-411f-81dd-e2d702d6dd8b", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.84", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "cc50c08b679248798f76b07c1530ed7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "503991c4-44d0-42d9-aa03-5259331f1051", "external-id": "nsx-vlan-transportzone-3", "segmentation_id": 3, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap664021c6-82", "ovs_interfaceid": "664021c6-82fb-46ea-a7d2-c9e942e0dded", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 885.971114] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3d:ca:1e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '503991c4-44d0-42d9-aa03-5259331f1051', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '664021c6-82fb-46ea-a7d2-c9e942e0dded', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 885.979384] env[63021]: DEBUG oslo.service.loopingcall [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 885.980091] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 885.980377] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9073893d-1d50-4e8e-87d5-290245b8175f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.003840] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 886.003840] env[63021]: value = "task-1293845" [ 886.003840] env[63021]: _type = "Task" [ 886.003840] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.014302] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293845, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.129280] env[63021]: DEBUG nova.scheduler.client.report [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 886.155490] env[63021]: DEBUG nova.compute.manager [req-8f664fab-c665-48f9-b405-f4734137fd50 req-98bedffa-10a5-42ff-969e-81a0dc393557 service nova] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Received event network-changed-664021c6-82fb-46ea-a7d2-c9e942e0dded {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 886.155821] env[63021]: DEBUG nova.compute.manager [req-8f664fab-c665-48f9-b405-f4734137fd50 req-98bedffa-10a5-42ff-969e-81a0dc393557 service nova] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Refreshing instance network info cache due to event network-changed-664021c6-82fb-46ea-a7d2-c9e942e0dded. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 886.155919] env[63021]: DEBUG oslo_concurrency.lockutils [req-8f664fab-c665-48f9-b405-f4734137fd50 req-98bedffa-10a5-42ff-969e-81a0dc393557 service nova] Acquiring lock "refresh_cache-51e4a14c-4792-43cd-9e99-4e06a8f6d6fb" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.156086] env[63021]: DEBUG oslo_concurrency.lockutils [req-8f664fab-c665-48f9-b405-f4734137fd50 req-98bedffa-10a5-42ff-969e-81a0dc393557 service nova] Acquired lock "refresh_cache-51e4a14c-4792-43cd-9e99-4e06a8f6d6fb" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.156234] env[63021]: DEBUG nova.network.neutron [req-8f664fab-c665-48f9-b405-f4734137fd50 req-98bedffa-10a5-42ff-969e-81a0dc393557 service nova] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Refreshing network info cache for port 664021c6-82fb-46ea-a7d2-c9e942e0dded {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 886.175650] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293844, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.230587] env[63021]: DEBUG oslo_vmware.api [None req-dcf93f26-6ad3-471b-b85c-2f39720207f0 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293837, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.265937] env[63021]: DEBUG nova.compute.manager [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 886.287762] env[63021]: DEBUG nova.virt.hardware [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 886.288023] env[63021]: DEBUG nova.virt.hardware [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 886.288198] env[63021]: DEBUG nova.virt.hardware [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 886.289027] env[63021]: DEBUG nova.virt.hardware [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 886.289027] env[63021]: DEBUG nova.virt.hardware [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 886.289027] env[63021]: DEBUG nova.virt.hardware [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 886.289027] env[63021]: DEBUG nova.virt.hardware [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 886.289027] env[63021]: DEBUG nova.virt.hardware [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 886.289234] env[63021]: DEBUG nova.virt.hardware [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 886.289334] env[63021]: DEBUG nova.virt.hardware [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 886.289494] env[63021]: DEBUG nova.virt.hardware [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 886.290800] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5253734d-ee89-4fc0-bef4-73b11951ed30 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.299310] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8a35d76-a189-46e1-bb33-1635e0821b64 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.441874] env[63021]: INFO nova.compute.manager [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Took 38.70 seconds to build instance. [ 886.514839] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293845, 'name': CreateVM_Task} progress is 25%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.637020] env[63021]: DEBUG oslo_concurrency.lockutils [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.396s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.637020] env[63021]: DEBUG nova.compute.manager [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 886.640636] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.623s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.640911] env[63021]: DEBUG nova.objects.instance [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Lazy-loading 'resources' on Instance uuid 70f67905-7c60-433d-9ebc-d66fa44eb36e {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 886.677328] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293844, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.583116} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.677633] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed/50a1269e-4d83-4cf7-ba14-3b1afb3c43ed.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 886.677872] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 886.678378] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f94ff82b-9892-436a-8877-bf9bb2d275c9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.687024] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for the task: (returnval){ [ 886.687024] env[63021]: value = "task-1293846" [ 886.687024] env[63021]: _type = "Task" [ 886.687024] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.696915] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293846, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.725327] env[63021]: DEBUG nova.network.neutron [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Successfully updated port: d34cfd0d-464e-400b-8366-83321cc6112f {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 886.730076] env[63021]: DEBUG oslo_vmware.api [None req-dcf93f26-6ad3-471b-b85c-2f39720207f0 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293837, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.909154] env[63021]: DEBUG nova.network.neutron [req-8f664fab-c665-48f9-b405-f4734137fd50 req-98bedffa-10a5-42ff-969e-81a0dc393557 service nova] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Updated VIF entry in instance network info cache for port 664021c6-82fb-46ea-a7d2-c9e942e0dded. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 886.909892] env[63021]: DEBUG nova.network.neutron [req-8f664fab-c665-48f9-b405-f4734137fd50 req-98bedffa-10a5-42ff-969e-81a0dc393557 service nova] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Updating instance_info_cache with network_info: [{"id": "664021c6-82fb-46ea-a7d2-c9e942e0dded", "address": "fa:16:3e:3d:ca:1e", "network": {"id": "d67a9f2c-02da-411f-81dd-e2d702d6dd8b", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.84", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "cc50c08b679248798f76b07c1530ed7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "503991c4-44d0-42d9-aa03-5259331f1051", "external-id": "nsx-vlan-transportzone-3", "segmentation_id": 3, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap664021c6-82", "ovs_interfaceid": "664021c6-82fb-46ea-a7d2-c9e942e0dded", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.944499] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Lock "44dadf8e-a727-448f-887d-6408f7ebfbc6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.500s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.016202] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293845, 'name': CreateVM_Task, 'duration_secs': 0.756967} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.016361] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 887.017259] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.017452] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.017782] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 887.018137] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff6416bc-4ba4-4f29-a7d7-ece13a7e42c9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.023141] env[63021]: DEBUG oslo_vmware.api [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Waiting for the task: (returnval){ [ 887.023141] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5269c11d-d128-60e1-fc53-9267bf5e4582" [ 887.023141] env[63021]: _type = "Task" [ 887.023141] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.032110] env[63021]: DEBUG oslo_vmware.api [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5269c11d-d128-60e1-fc53-9267bf5e4582, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.142580] env[63021]: DEBUG nova.compute.utils [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 887.143998] env[63021]: DEBUG nova.compute.manager [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 887.144177] env[63021]: DEBUG nova.network.neutron [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 887.191403] env[63021]: DEBUG nova.policy [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2e78e22dcacf41308c85ef1a0704e1f9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '52e7242ba61e4ce28c8f3b792980bdac', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 887.201226] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293846, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087148} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.201509] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 887.205236] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ce0d84e-b425-4ad9-ba53-6af94c1f0179 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.227048] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed/50a1269e-4d83-4cf7-ba14-3b1afb3c43ed.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 887.231699] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9a54ae4d-5a65-493f-b9a7-0cab97cd0160 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.246483] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquiring lock "refresh_cache-1b4871cc-3dac-4683-8fce-93c4d9e4407d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.246632] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquired lock "refresh_cache-1b4871cc-3dac-4683-8fce-93c4d9e4407d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.246816] env[63021]: DEBUG nova.network.neutron [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 887.257740] env[63021]: DEBUG oslo_vmware.api [None req-dcf93f26-6ad3-471b-b85c-2f39720207f0 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293837, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.262923] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for the task: (returnval){ [ 887.262923] env[63021]: value = "task-1293847" [ 887.262923] env[63021]: _type = "Task" [ 887.262923] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.272790] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293847, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.416028] env[63021]: DEBUG oslo_concurrency.lockutils [req-8f664fab-c665-48f9-b405-f4734137fd50 req-98bedffa-10a5-42ff-969e-81a0dc393557 service nova] Releasing lock "refresh_cache-51e4a14c-4792-43cd-9e99-4e06a8f6d6fb" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.495358] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-304389c5-0abf-45fc-a826-a8578052644a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.500737] env[63021]: DEBUG nova.network.neutron [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Successfully created port: 3dc5a6b8-23fe-42fa-88c1-6854a326150d {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 887.505918] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e513628-2581-4ce5-9e3a-fa6a3b2d1ec7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.543767] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4b19127-6af8-4511-a409-8113cf557712 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.552356] env[63021]: DEBUG oslo_vmware.api [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5269c11d-d128-60e1-fc53-9267bf5e4582, 'name': SearchDatastore_Task, 'duration_secs': 0.011238} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.554330] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.554632] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 887.554901] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.555095] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.555292] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 887.555677] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0489ac53-2555-4c97-b842-587f33097143 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.558651] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b71b3621-cf26-42d9-b355-75b7727ebc08 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.576157] env[63021]: DEBUG nova.compute.provider_tree [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 887.576157] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 887.576322] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 887.577411] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78e2f924-47fd-443e-b7c4-8c1f061bfdd0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.583165] env[63021]: DEBUG oslo_vmware.api [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Waiting for the task: (returnval){ [ 887.583165] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5237cf20-a195-4718-2d5e-2bb21fd1e4de" [ 887.583165] env[63021]: _type = "Task" [ 887.583165] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.592552] env[63021]: DEBUG oslo_vmware.api [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5237cf20-a195-4718-2d5e-2bb21fd1e4de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.649915] env[63021]: DEBUG nova.compute.manager [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 887.742269] env[63021]: DEBUG oslo_vmware.api [None req-dcf93f26-6ad3-471b-b85c-2f39720207f0 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293837, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.777178] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293847, 'name': ReconfigVM_Task, 'duration_secs': 0.445505} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.777478] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Reconfigured VM instance instance-0000004a to attach disk [datastore1] 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed/50a1269e-4d83-4cf7-ba14-3b1afb3c43ed.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 887.778153] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2151261e-1333-4860-9e41-aea3b074c2fb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.784773] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for the task: (returnval){ [ 887.784773] env[63021]: value = "task-1293848" [ 887.784773] env[63021]: _type = "Task" [ 887.784773] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.794502] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293848, 'name': Rename_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.802222] env[63021]: DEBUG nova.network.neutron [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 887.921413] env[63021]: DEBUG nova.network.neutron [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Updating instance_info_cache with network_info: [{"id": "d34cfd0d-464e-400b-8366-83321cc6112f", "address": "fa:16:3e:64:67:d0", "network": {"id": "37008b1f-68ff-404c-9afc-14b108af85c3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1473182647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "2d364c9eae4b4f8a8b60d4a3ac2ed648", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "02bbcead-d833-4543-bec6-fb82dfe659ff", "external-id": "nsx-vlan-transportzone-478", "segmentation_id": 478, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd34cfd0d-46", "ovs_interfaceid": "d34cfd0d-464e-400b-8366-83321cc6112f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.078928] env[63021]: DEBUG nova.scheduler.client.report [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 888.094447] env[63021]: DEBUG oslo_vmware.api [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5237cf20-a195-4718-2d5e-2bb21fd1e4de, 'name': SearchDatastore_Task, 'duration_secs': 0.014034} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.094447] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60830b28-79e3-4a58-876c-56efe548e875 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.100885] env[63021]: DEBUG oslo_vmware.api [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Waiting for the task: (returnval){ [ 888.100885] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]524760c0-bad9-5c22-1970-39e62696afa1" [ 888.100885] env[63021]: _type = "Task" [ 888.100885] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.111261] env[63021]: DEBUG oslo_vmware.api [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]524760c0-bad9-5c22-1970-39e62696afa1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.158419] env[63021]: INFO nova.virt.block_device [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Booting with volume e424adc5-a8b0-4409-b3e0-f3cdbc8fb712 at /dev/sda [ 888.196998] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-75d0e2e8-5a5a-413c-86b4-92e297f37b84 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.209774] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91e2e782-2f53-48da-b665-307fe1151c95 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.252512] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bdc34710-1c3d-4e6f-89eb-e3ac042275e3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.257724] env[63021]: DEBUG oslo_vmware.api [None req-dcf93f26-6ad3-471b-b85c-2f39720207f0 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293837, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.264255] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ad1b80f-a5b4-4ae1-86a5-5382dc60799e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.278490] env[63021]: DEBUG nova.compute.manager [req-e3076679-4047-4b9e-b97b-aa291033ae31 req-e5f578d3-74b8-42a1-b392-be9092966575 service nova] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Received event network-vif-plugged-d34cfd0d-464e-400b-8366-83321cc6112f {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 888.278792] env[63021]: DEBUG oslo_concurrency.lockutils [req-e3076679-4047-4b9e-b97b-aa291033ae31 req-e5f578d3-74b8-42a1-b392-be9092966575 service nova] Acquiring lock "1b4871cc-3dac-4683-8fce-93c4d9e4407d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.278914] env[63021]: DEBUG oslo_concurrency.lockutils [req-e3076679-4047-4b9e-b97b-aa291033ae31 req-e5f578d3-74b8-42a1-b392-be9092966575 service nova] Lock "1b4871cc-3dac-4683-8fce-93c4d9e4407d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.279093] env[63021]: DEBUG oslo_concurrency.lockutils [req-e3076679-4047-4b9e-b97b-aa291033ae31 req-e5f578d3-74b8-42a1-b392-be9092966575 service nova] Lock "1b4871cc-3dac-4683-8fce-93c4d9e4407d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.279262] env[63021]: DEBUG nova.compute.manager [req-e3076679-4047-4b9e-b97b-aa291033ae31 req-e5f578d3-74b8-42a1-b392-be9092966575 service nova] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] No waiting events found dispatching network-vif-plugged-d34cfd0d-464e-400b-8366-83321cc6112f {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 888.279557] env[63021]: WARNING nova.compute.manager [req-e3076679-4047-4b9e-b97b-aa291033ae31 req-e5f578d3-74b8-42a1-b392-be9092966575 service nova] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Received unexpected event network-vif-plugged-d34cfd0d-464e-400b-8366-83321cc6112f for instance with vm_state building and task_state spawning. [ 888.279731] env[63021]: DEBUG nova.compute.manager [req-e3076679-4047-4b9e-b97b-aa291033ae31 req-e5f578d3-74b8-42a1-b392-be9092966575 service nova] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Received event network-changed-d34cfd0d-464e-400b-8366-83321cc6112f {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 888.279890] env[63021]: DEBUG nova.compute.manager [req-e3076679-4047-4b9e-b97b-aa291033ae31 req-e5f578d3-74b8-42a1-b392-be9092966575 service nova] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Refreshing instance network info cache due to event network-changed-d34cfd0d-464e-400b-8366-83321cc6112f. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 888.280134] env[63021]: DEBUG oslo_concurrency.lockutils [req-e3076679-4047-4b9e-b97b-aa291033ae31 req-e5f578d3-74b8-42a1-b392-be9092966575 service nova] Acquiring lock "refresh_cache-1b4871cc-3dac-4683-8fce-93c4d9e4407d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.305296] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4cc94ad-1d98-4b4a-aa85-0cd8ada8c184 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.309485] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293848, 'name': Rename_Task, 'duration_secs': 0.218697} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.309485] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 888.309485] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-47d28bee-11a5-455d-b271-4988441429a6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.313208] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0026fe5c-625c-4d9c-81e8-6295145b2afd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.316860] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for the task: (returnval){ [ 888.316860] env[63021]: value = "task-1293849" [ 888.316860] env[63021]: _type = "Task" [ 888.316860] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.328783] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293849, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.331966] env[63021]: DEBUG nova.virt.block_device [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Updating existing volume attachment record: 3fd0f721-b6c5-4dba-9df0-2b46964236a6 {{(pid=63021) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 888.425284] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Releasing lock "refresh_cache-1b4871cc-3dac-4683-8fce-93c4d9e4407d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.425284] env[63021]: DEBUG nova.compute.manager [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Instance network_info: |[{"id": "d34cfd0d-464e-400b-8366-83321cc6112f", "address": "fa:16:3e:64:67:d0", "network": {"id": "37008b1f-68ff-404c-9afc-14b108af85c3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1473182647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "2d364c9eae4b4f8a8b60d4a3ac2ed648", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "02bbcead-d833-4543-bec6-fb82dfe659ff", "external-id": "nsx-vlan-transportzone-478", "segmentation_id": 478, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd34cfd0d-46", "ovs_interfaceid": "d34cfd0d-464e-400b-8366-83321cc6112f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 888.425284] env[63021]: DEBUG oslo_concurrency.lockutils [req-e3076679-4047-4b9e-b97b-aa291033ae31 req-e5f578d3-74b8-42a1-b392-be9092966575 service nova] Acquired lock "refresh_cache-1b4871cc-3dac-4683-8fce-93c4d9e4407d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.425284] env[63021]: DEBUG nova.network.neutron [req-e3076679-4047-4b9e-b97b-aa291033ae31 req-e5f578d3-74b8-42a1-b392-be9092966575 service nova] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Refreshing network info cache for port d34cfd0d-464e-400b-8366-83321cc6112f {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 888.426563] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:64:67:d0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '02bbcead-d833-4543-bec6-fb82dfe659ff', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd34cfd0d-464e-400b-8366-83321cc6112f', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 888.433928] env[63021]: DEBUG oslo.service.loopingcall [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 888.435055] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 888.435317] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-20c15598-f8eb-44af-b359-f27b24d7b525 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.455296] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 888.455296] env[63021]: value = "task-1293850" [ 888.455296] env[63021]: _type = "Task" [ 888.455296] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.463812] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293850, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.584475] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.944s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.587788] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 26.779s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.611920] env[63021]: DEBUG oslo_vmware.api [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]524760c0-bad9-5c22-1970-39e62696afa1, 'name': SearchDatastore_Task, 'duration_secs': 0.012913} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.612954] env[63021]: INFO nova.scheduler.client.report [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Deleted allocations for instance 70f67905-7c60-433d-9ebc-d66fa44eb36e [ 888.613951] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.614228] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb/51e4a14c-4792-43cd-9e99-4e06a8f6d6fb.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 888.616946] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-56d8fd30-072a-457b-90a4-7db9d593297c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.625278] env[63021]: DEBUG oslo_vmware.api [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Waiting for the task: (returnval){ [ 888.625278] env[63021]: value = "task-1293851" [ 888.625278] env[63021]: _type = "Task" [ 888.625278] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.633865] env[63021]: DEBUG oslo_vmware.api [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293851, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.742799] env[63021]: DEBUG oslo_vmware.api [None req-dcf93f26-6ad3-471b-b85c-2f39720207f0 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293837, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.828521] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293849, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.971123] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293850, 'name': CreateVM_Task} progress is 25%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.100049] env[63021]: INFO nova.compute.claims [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 889.124009] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e120f8a2-afef-4ff6-b780-aee5fa0e73b0 tempest-ServerGroupTestJSON-1234736076 tempest-ServerGroupTestJSON-1234736076-project-member] Lock "70f67905-7c60-433d-9ebc-d66fa44eb36e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.004s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.152637] env[63021]: DEBUG oslo_vmware.api [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293851, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.238071] env[63021]: DEBUG nova.network.neutron [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Successfully updated port: 3dc5a6b8-23fe-42fa-88c1-6854a326150d {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 889.245064] env[63021]: DEBUG oslo_vmware.api [None req-dcf93f26-6ad3-471b-b85c-2f39720207f0 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293837, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.293088] env[63021]: DEBUG nova.network.neutron [req-e3076679-4047-4b9e-b97b-aa291033ae31 req-e5f578d3-74b8-42a1-b392-be9092966575 service nova] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Updated VIF entry in instance network info cache for port d34cfd0d-464e-400b-8366-83321cc6112f. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 889.293452] env[63021]: DEBUG nova.network.neutron [req-e3076679-4047-4b9e-b97b-aa291033ae31 req-e5f578d3-74b8-42a1-b392-be9092966575 service nova] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Updating instance_info_cache with network_info: [{"id": "d34cfd0d-464e-400b-8366-83321cc6112f", "address": "fa:16:3e:64:67:d0", "network": {"id": "37008b1f-68ff-404c-9afc-14b108af85c3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1473182647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "2d364c9eae4b4f8a8b60d4a3ac2ed648", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "02bbcead-d833-4543-bec6-fb82dfe659ff", "external-id": "nsx-vlan-transportzone-478", "segmentation_id": 478, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd34cfd0d-46", "ovs_interfaceid": "d34cfd0d-464e-400b-8366-83321cc6112f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.329334] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293849, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.466022] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293850, 'name': CreateVM_Task} progress is 25%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.609282] env[63021]: INFO nova.compute.resource_tracker [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Updating resource usage from migration b36d8c79-3341-4688-9479-6846fc53dcad [ 889.640936] env[63021]: DEBUG oslo_vmware.api [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293851, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.556018} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.641161] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb/51e4a14c-4792-43cd-9e99-4e06a8f6d6fb.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 889.641384] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 889.641727] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4d8408fa-8869-4105-9189-e59b82a191ba {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.649064] env[63021]: DEBUG oslo_vmware.api [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Waiting for the task: (returnval){ [ 889.649064] env[63021]: value = "task-1293852" [ 889.649064] env[63021]: _type = "Task" [ 889.649064] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.660016] env[63021]: DEBUG oslo_vmware.api [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293852, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.741115] env[63021]: DEBUG oslo_concurrency.lockutils [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Acquiring lock "refresh_cache-80b857c4-d9dd-4483-970f-a0f5f4bcf173" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.741115] env[63021]: DEBUG oslo_concurrency.lockutils [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Acquired lock "refresh_cache-80b857c4-d9dd-4483-970f-a0f5f4bcf173" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.741115] env[63021]: DEBUG nova.network.neutron [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 889.751979] env[63021]: DEBUG oslo_vmware.api [None req-dcf93f26-6ad3-471b-b85c-2f39720207f0 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293837, 'name': ReconfigVM_Task, 'duration_secs': 6.303147} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.751979] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dcf93f26-6ad3-471b-b85c-2f39720207f0 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Releasing lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.751979] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-dcf93f26-6ad3-471b-b85c-2f39720207f0 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Reconfigured VM to detach interface {{(pid=63021) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 889.795608] env[63021]: DEBUG oslo_concurrency.lockutils [req-e3076679-4047-4b9e-b97b-aa291033ae31 req-e5f578d3-74b8-42a1-b392-be9092966575 service nova] Releasing lock "refresh_cache-1b4871cc-3dac-4683-8fce-93c4d9e4407d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.830371] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293849, 'name': PowerOnVM_Task} progress is 78%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.950871] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11c7ecba-46b7-4d7e-b87b-b7e7e00f439e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.962327] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-866f2407-c121-4e81-8db3-5f8d7d9c92bc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.972510] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293850, 'name': CreateVM_Task, 'duration_secs': 1.300357} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.010925] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 890.012580] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.012771] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.013156] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 890.014065] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a69ad30c-da49-4213-98a8-edb875613bf9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.017279] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7490a3fd-0e38-46bc-ac23-c86b4155aeb7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.022864] env[63021]: DEBUG oslo_vmware.api [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 890.022864] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]529a87d1-fd08-8f84-a33e-35502094fa55" [ 890.022864] env[63021]: _type = "Task" [ 890.022864] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.029608] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a589763-d322-476b-a9a1-fbc12592d806 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.039697] env[63021]: DEBUG oslo_vmware.api [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]529a87d1-fd08-8f84-a33e-35502094fa55, 'name': SearchDatastore_Task, 'duration_secs': 0.014428} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.047597] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.047852] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 890.049237] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.049237] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.049237] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 890.049237] env[63021]: DEBUG nova.compute.provider_tree [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.050161] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2b3f0a0c-1273-4044-ba39-3eda08f8206e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.060081] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 890.060285] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 890.061079] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86e3984e-26a6-42cd-a67b-1174bc4e1f22 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.067368] env[63021]: DEBUG oslo_vmware.api [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 890.067368] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52879d40-d8d2-51fd-f6f2-09214b679973" [ 890.067368] env[63021]: _type = "Task" [ 890.067368] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.075678] env[63021]: DEBUG oslo_vmware.api [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52879d40-d8d2-51fd-f6f2-09214b679973, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.158927] env[63021]: DEBUG oslo_vmware.api [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293852, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.106491} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.163425] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 890.164428] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7afa7620-d0dc-42ec-8128-207d67f72cfa {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.188379] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb/51e4a14c-4792-43cd-9e99-4e06a8f6d6fb.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 890.190018] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8ec58c1e-d8b2-43f4-b2a9-eb0a186be870 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.211249] env[63021]: DEBUG oslo_vmware.api [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Waiting for the task: (returnval){ [ 890.211249] env[63021]: value = "task-1293853" [ 890.211249] env[63021]: _type = "Task" [ 890.211249] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.220759] env[63021]: DEBUG oslo_vmware.api [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293853, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.310468] env[63021]: DEBUG nova.network.neutron [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 890.316068] env[63021]: DEBUG nova.compute.manager [req-227023fa-159b-4e6b-9bd5-4071fb5a5e92 req-dc68025a-81f9-4788-940e-bca9bd14141f service nova] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Received event network-vif-plugged-3dc5a6b8-23fe-42fa-88c1-6854a326150d {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 890.316341] env[63021]: DEBUG oslo_concurrency.lockutils [req-227023fa-159b-4e6b-9bd5-4071fb5a5e92 req-dc68025a-81f9-4788-940e-bca9bd14141f service nova] Acquiring lock "80b857c4-d9dd-4483-970f-a0f5f4bcf173-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.316649] env[63021]: DEBUG oslo_concurrency.lockutils [req-227023fa-159b-4e6b-9bd5-4071fb5a5e92 req-dc68025a-81f9-4788-940e-bca9bd14141f service nova] Lock "80b857c4-d9dd-4483-970f-a0f5f4bcf173-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.316934] env[63021]: DEBUG oslo_concurrency.lockutils [req-227023fa-159b-4e6b-9bd5-4071fb5a5e92 req-dc68025a-81f9-4788-940e-bca9bd14141f service nova] Lock "80b857c4-d9dd-4483-970f-a0f5f4bcf173-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.317153] env[63021]: DEBUG nova.compute.manager [req-227023fa-159b-4e6b-9bd5-4071fb5a5e92 req-dc68025a-81f9-4788-940e-bca9bd14141f service nova] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] No waiting events found dispatching network-vif-plugged-3dc5a6b8-23fe-42fa-88c1-6854a326150d {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 890.317386] env[63021]: WARNING nova.compute.manager [req-227023fa-159b-4e6b-9bd5-4071fb5a5e92 req-dc68025a-81f9-4788-940e-bca9bd14141f service nova] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Received unexpected event network-vif-plugged-3dc5a6b8-23fe-42fa-88c1-6854a326150d for instance with vm_state building and task_state spawning. [ 890.317608] env[63021]: DEBUG nova.compute.manager [req-227023fa-159b-4e6b-9bd5-4071fb5a5e92 req-dc68025a-81f9-4788-940e-bca9bd14141f service nova] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Received event network-changed-3dc5a6b8-23fe-42fa-88c1-6854a326150d {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 890.317856] env[63021]: DEBUG nova.compute.manager [req-227023fa-159b-4e6b-9bd5-4071fb5a5e92 req-dc68025a-81f9-4788-940e-bca9bd14141f service nova] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Refreshing instance network info cache due to event network-changed-3dc5a6b8-23fe-42fa-88c1-6854a326150d. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 890.318120] env[63021]: DEBUG oslo_concurrency.lockutils [req-227023fa-159b-4e6b-9bd5-4071fb5a5e92 req-dc68025a-81f9-4788-940e-bca9bd14141f service nova] Acquiring lock "refresh_cache-80b857c4-d9dd-4483-970f-a0f5f4bcf173" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.332188] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293849, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.424144] env[63021]: DEBUG nova.compute.manager [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 890.424793] env[63021]: DEBUG nova.virt.hardware [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 890.425072] env[63021]: DEBUG nova.virt.hardware [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 890.425245] env[63021]: DEBUG nova.virt.hardware [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 890.425430] env[63021]: DEBUG nova.virt.hardware [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 890.425575] env[63021]: DEBUG nova.virt.hardware [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 890.425916] env[63021]: DEBUG nova.virt.hardware [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 890.425916] env[63021]: DEBUG nova.virt.hardware [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 890.426096] env[63021]: DEBUG nova.virt.hardware [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 890.426284] env[63021]: DEBUG nova.virt.hardware [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 890.426498] env[63021]: DEBUG nova.virt.hardware [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 890.426733] env[63021]: DEBUG nova.virt.hardware [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 890.427982] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fd78a46-ac5e-4476-acdf-cd3a97dc3c11 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.438280] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dab5536-fc21-4134-9c28-3bfd20dd5ff9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.557020] env[63021]: DEBUG nova.scheduler.client.report [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 890.582026] env[63021]: DEBUG oslo_vmware.api [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52879d40-d8d2-51fd-f6f2-09214b679973, 'name': SearchDatastore_Task, 'duration_secs': 0.020106} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.582026] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5891d4c5-211b-4d8b-82dc-8391f84359ee {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.586801] env[63021]: DEBUG nova.network.neutron [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Updating instance_info_cache with network_info: [{"id": "3dc5a6b8-23fe-42fa-88c1-6854a326150d", "address": "fa:16:3e:13:6a:e4", "network": {"id": "1a4029b0-70bb-49c1-9689-7517cf4b29d3", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-875049177-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52e7242ba61e4ce28c8f3b792980bdac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f762954-6ca5-4da5-bf0a-5d31c51ec570", "external-id": "nsx-vlan-transportzone-930", "segmentation_id": 930, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3dc5a6b8-23", "ovs_interfaceid": "3dc5a6b8-23fe-42fa-88c1-6854a326150d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.594025] env[63021]: DEBUG oslo_vmware.api [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 890.594025] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52429efa-78ea-a8fd-9041-ab2c4f9885ba" [ 890.594025] env[63021]: _type = "Task" [ 890.594025] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.602156] env[63021]: DEBUG oslo_vmware.api [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52429efa-78ea-a8fd-9041-ab2c4f9885ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.710543] env[63021]: DEBUG nova.compute.manager [req-492ee330-42de-456c-ae0f-7b6b5035ee63 req-8ea74433-628f-4d15-9d29-720aabfc940c service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Received event network-vif-deleted-2f88c221-8f47-478e-a5f3-ba0a230c7ad4 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 890.713347] env[63021]: INFO nova.compute.manager [req-492ee330-42de-456c-ae0f-7b6b5035ee63 req-8ea74433-628f-4d15-9d29-720aabfc940c service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Neutron deleted interface 2f88c221-8f47-478e-a5f3-ba0a230c7ad4; detaching it from the instance and deleting it from the info cache [ 890.713347] env[63021]: DEBUG nova.network.neutron [req-492ee330-42de-456c-ae0f-7b6b5035ee63 req-8ea74433-628f-4d15-9d29-720aabfc940c service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Updating instance_info_cache with network_info: [{"id": "37dcb3f1-2931-40c6-96ba-b31b8fa1a72b", "address": "fa:16:3e:57:13:17", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37dcb3f1-29", "ovs_interfaceid": "37dcb3f1-2931-40c6-96ba-b31b8fa1a72b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d92edf1f-d15d-4f56-9da3-92cbe44b6709", "address": "fa:16:3e:b5:1b:c8", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd92edf1f-d1", "ovs_interfaceid": "d92edf1f-d15d-4f56-9da3-92cbe44b6709", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.729773] env[63021]: DEBUG oslo_vmware.api [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293853, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.836195] env[63021]: DEBUG oslo_vmware.api [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293849, 'name': PowerOnVM_Task, 'duration_secs': 2.126876} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.836195] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 890.836195] env[63021]: INFO nova.compute.manager [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Took 10.59 seconds to spawn the instance on the hypervisor. [ 890.836195] env[63021]: DEBUG nova.compute.manager [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 890.836195] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c28b981-f3d5-4500-878e-45bf5ec4caa9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.062979] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.475s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.063362] env[63021]: INFO nova.compute.manager [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Migrating [ 891.063650] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.063944] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquired lock "compute-rpcapi-router" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.068808] env[63021]: DEBUG oslo_concurrency.lockutils [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.920s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.068808] env[63021]: DEBUG nova.objects.instance [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Lazy-loading 'resources' on Instance uuid 6f4e6245-c0d9-46b4-baf5-5aca19e66da5 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 891.090237] env[63021]: DEBUG oslo_concurrency.lockutils [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Releasing lock "refresh_cache-80b857c4-d9dd-4483-970f-a0f5f4bcf173" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.090591] env[63021]: DEBUG nova.compute.manager [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Instance network_info: |[{"id": "3dc5a6b8-23fe-42fa-88c1-6854a326150d", "address": "fa:16:3e:13:6a:e4", "network": {"id": "1a4029b0-70bb-49c1-9689-7517cf4b29d3", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-875049177-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52e7242ba61e4ce28c8f3b792980bdac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f762954-6ca5-4da5-bf0a-5d31c51ec570", "external-id": "nsx-vlan-transportzone-930", "segmentation_id": 930, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3dc5a6b8-23", "ovs_interfaceid": "3dc5a6b8-23fe-42fa-88c1-6854a326150d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 891.091225] env[63021]: DEBUG oslo_concurrency.lockutils [req-227023fa-159b-4e6b-9bd5-4071fb5a5e92 req-dc68025a-81f9-4788-940e-bca9bd14141f service nova] Acquired lock "refresh_cache-80b857c4-d9dd-4483-970f-a0f5f4bcf173" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.091417] env[63021]: DEBUG nova.network.neutron [req-227023fa-159b-4e6b-9bd5-4071fb5a5e92 req-dc68025a-81f9-4788-940e-bca9bd14141f service nova] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Refreshing network info cache for port 3dc5a6b8-23fe-42fa-88c1-6854a326150d {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 891.100283] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:13:6a:e4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1f762954-6ca5-4da5-bf0a-5d31c51ec570', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3dc5a6b8-23fe-42fa-88c1-6854a326150d', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 891.117378] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Creating folder: Project (52e7242ba61e4ce28c8f3b792980bdac). Parent ref: group-v277447. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 891.119128] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5cdc3de3-3530-4895-b540-3025550befb4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.134717] env[63021]: DEBUG oslo_vmware.api [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52429efa-78ea-a8fd-9041-ab2c4f9885ba, 'name': SearchDatastore_Task, 'duration_secs': 0.017913} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.134717] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.134717] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 1b4871cc-3dac-4683-8fce-93c4d9e4407d/1b4871cc-3dac-4683-8fce-93c4d9e4407d.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 891.134717] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-16bec9ae-5e34-4f74-8200-d688f36664b2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.136892] env[63021]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 891.137064] env[63021]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=63021) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 891.137761] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Folder already exists: Project (52e7242ba61e4ce28c8f3b792980bdac). Parent ref: group-v277447. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 891.138026] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Creating folder: Instances. Parent ref: group-v277489. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 891.138704] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6b929364-1df0-48db-9dfa-1cd6c088fe4e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.142805] env[63021]: DEBUG oslo_vmware.api [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 891.142805] env[63021]: value = "task-1293855" [ 891.142805] env[63021]: _type = "Task" [ 891.142805] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.149188] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Created folder: Instances in parent group-v277489. [ 891.149188] env[63021]: DEBUG oslo.service.loopingcall [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 891.151637] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 891.151930] env[63021]: DEBUG oslo_vmware.api [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293855, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.152684] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ba287c02-5046-4389-af8d-049bfc2523a3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.172331] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 891.172331] env[63021]: value = "task-1293857" [ 891.172331] env[63021]: _type = "Task" [ 891.172331] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.180558] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293857, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.215907] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dcf93f26-6ad3-471b-b85c-2f39720207f0 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "refresh_cache-36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.216127] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dcf93f26-6ad3-471b-b85c-2f39720207f0 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquired lock "refresh_cache-36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.216235] env[63021]: DEBUG nova.network.neutron [None req-dcf93f26-6ad3-471b-b85c-2f39720207f0 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 891.221018] env[63021]: DEBUG oslo_concurrency.lockutils [req-492ee330-42de-456c-ae0f-7b6b5035ee63 req-8ea74433-628f-4d15-9d29-720aabfc940c service nova] Acquiring lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.221018] env[63021]: DEBUG oslo_concurrency.lockutils [req-492ee330-42de-456c-ae0f-7b6b5035ee63 req-8ea74433-628f-4d15-9d29-720aabfc940c service nova] Acquired lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.222370] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79666468-e5e9-4f02-97dd-9516db7b675a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.233283] env[63021]: DEBUG oslo_vmware.api [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293853, 'name': ReconfigVM_Task, 'duration_secs': 0.762125} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.247131] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Reconfigured VM instance instance-0000004b to attach disk [datastore2] 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb/51e4a14c-4792-43cd-9e99-4e06a8f6d6fb.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 891.248712] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f1ea84e9-fa4b-4dd2-a2bb-9c3f7c155da4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.250602] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeaa4e71-6705-40ea-ade7-75f9bb5201d8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.272886] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.281716] env[63021]: DEBUG nova.virt.vmwareapi.vmops [req-492ee330-42de-456c-ae0f-7b6b5035ee63 req-8ea74433-628f-4d15-9d29-720aabfc940c service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Reconfiguring VM to detach interface {{(pid=63021) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 891.281716] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3dead8df-de66-4226-9c55-14ec2d4fb8d6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.294588] env[63021]: DEBUG oslo_vmware.api [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Waiting for the task: (returnval){ [ 891.294588] env[63021]: value = "task-1293858" [ 891.294588] env[63021]: _type = "Task" [ 891.294588] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.299660] env[63021]: DEBUG oslo_vmware.api [req-492ee330-42de-456c-ae0f-7b6b5035ee63 req-8ea74433-628f-4d15-9d29-720aabfc940c service nova] Waiting for the task: (returnval){ [ 891.299660] env[63021]: value = "task-1293859" [ 891.299660] env[63021]: _type = "Task" [ 891.299660] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.311123] env[63021]: DEBUG oslo_vmware.api [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293858, 'name': Rename_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.316736] env[63021]: DEBUG oslo_vmware.api [req-492ee330-42de-456c-ae0f-7b6b5035ee63 req-8ea74433-628f-4d15-9d29-720aabfc940c service nova] Task: {'id': task-1293859, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.355287] env[63021]: INFO nova.compute.manager [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Took 40.16 seconds to build instance. [ 891.570517] env[63021]: INFO nova.compute.rpcapi [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 891.571376] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Releasing lock "compute-rpcapi-router" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.666360] env[63021]: DEBUG oslo_vmware.api [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293855, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.685862] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293857, 'name': CreateVM_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.813228] env[63021]: DEBUG oslo_vmware.api [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293858, 'name': Rename_Task, 'duration_secs': 0.222559} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.820603] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 891.825344] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bb458641-d697-4db8-ba9c-eea9c8624230 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.832906] env[63021]: DEBUG oslo_vmware.api [req-492ee330-42de-456c-ae0f-7b6b5035ee63 req-8ea74433-628f-4d15-9d29-720aabfc940c service nova] Task: {'id': task-1293859, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.841046] env[63021]: DEBUG oslo_vmware.api [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Waiting for the task: (returnval){ [ 891.841046] env[63021]: value = "task-1293860" [ 891.841046] env[63021]: _type = "Task" [ 891.841046] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.855539] env[63021]: DEBUG oslo_vmware.api [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293860, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.857409] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0e2edce4-bb97-4210-bf16-6e9136b4b61b tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Lock "50a1269e-4d83-4cf7-ba14-3b1afb3c43ed" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 86.379s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.909104] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 891.909421] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 891.988242] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9584fab-f342-4a32-ae3a-0ebb8e3c564f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.998179] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-417b55f9-7c60-4b9d-a00a-af0006d36f7f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.049828] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92654a8d-68b5-4510-9ff0-391528ca5fee {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.061224] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1927c708-3178-41f3-998a-04a9e10dda29 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.079498] env[63021]: DEBUG nova.compute.provider_tree [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 892.104354] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "refresh_cache-71909213-258d-45f5-9a3e-e473e8ce8aa2" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.104646] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquired lock "refresh_cache-71909213-258d-45f5-9a3e-e473e8ce8aa2" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.104933] env[63021]: DEBUG nova.network.neutron [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 892.138843] env[63021]: DEBUG nova.network.neutron [req-227023fa-159b-4e6b-9bd5-4071fb5a5e92 req-dc68025a-81f9-4788-940e-bca9bd14141f service nova] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Updated VIF entry in instance network info cache for port 3dc5a6b8-23fe-42fa-88c1-6854a326150d. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 892.139346] env[63021]: DEBUG nova.network.neutron [req-227023fa-159b-4e6b-9bd5-4071fb5a5e92 req-dc68025a-81f9-4788-940e-bca9bd14141f service nova] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Updating instance_info_cache with network_info: [{"id": "3dc5a6b8-23fe-42fa-88c1-6854a326150d", "address": "fa:16:3e:13:6a:e4", "network": {"id": "1a4029b0-70bb-49c1-9689-7517cf4b29d3", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-875049177-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52e7242ba61e4ce28c8f3b792980bdac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f762954-6ca5-4da5-bf0a-5d31c51ec570", "external-id": "nsx-vlan-transportzone-930", "segmentation_id": 930, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3dc5a6b8-23", "ovs_interfaceid": "3dc5a6b8-23fe-42fa-88c1-6854a326150d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.163638] env[63021]: DEBUG oslo_vmware.api [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293855, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.691824} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.164367] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 1b4871cc-3dac-4683-8fce-93c4d9e4407d/1b4871cc-3dac-4683-8fce-93c4d9e4407d.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 892.164854] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 892.165354] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5057008f-4029-4e8b-bd10-bb1c54cabbdc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.179549] env[63021]: DEBUG oslo_vmware.api [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 892.179549] env[63021]: value = "task-1293861" [ 892.179549] env[63021]: _type = "Task" [ 892.179549] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.187814] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293857, 'name': CreateVM_Task, 'duration_secs': 0.712958} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.188569] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 892.189825] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'delete_on_termination': True, 'mount_device': '/dev/sda', 'device_type': None, 'attachment_id': '3fd0f721-b6c5-4dba-9df0-2b46964236a6', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277493', 'volume_id': 'e424adc5-a8b0-4409-b3e0-f3cdbc8fb712', 'name': 'volume-e424adc5-a8b0-4409-b3e0-f3cdbc8fb712', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '80b857c4-d9dd-4483-970f-a0f5f4bcf173', 'attached_at': '', 'detached_at': '', 'volume_id': 'e424adc5-a8b0-4409-b3e0-f3cdbc8fb712', 'serial': 'e424adc5-a8b0-4409-b3e0-f3cdbc8fb712'}, 'guest_format': None, 'boot_index': 0, 'disk_bus': None, 'volume_type': None}], 'swap': None} {{(pid=63021) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 892.190296] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Root volume attach. Driver type: vmdk {{(pid=63021) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 892.191421] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-875462b1-1200-45a0-ac6b-b685e2b4f95d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.199670] env[63021]: DEBUG oslo_vmware.api [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293861, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.208209] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9581d3c2-915c-4829-aea9-68fc41b11fb3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.218944] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0042cd15-105d-4b2b-ba97-0b2c5739d230 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.226993] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-853bbfe3-f578-4b13-974c-40b65bb4fe66 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.236507] env[63021]: DEBUG oslo_vmware.api [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Waiting for the task: (returnval){ [ 892.236507] env[63021]: value = "task-1293862" [ 892.236507] env[63021]: _type = "Task" [ 892.236507] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.247546] env[63021]: DEBUG oslo_vmware.api [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Task: {'id': task-1293862, 'name': RelocateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.248676] env[63021]: INFO nova.network.neutron [None req-dcf93f26-6ad3-471b-b85c-2f39720207f0 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Port d92edf1f-d15d-4f56-9da3-92cbe44b6709 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 892.249290] env[63021]: DEBUG nova.network.neutron [None req-dcf93f26-6ad3-471b-b85c-2f39720207f0 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Updating instance_info_cache with network_info: [{"id": "37dcb3f1-2931-40c6-96ba-b31b8fa1a72b", "address": "fa:16:3e:57:13:17", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37dcb3f1-29", "ovs_interfaceid": "37dcb3f1-2931-40c6-96ba-b31b8fa1a72b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.316509] env[63021]: DEBUG oslo_vmware.api [req-492ee330-42de-456c-ae0f-7b6b5035ee63 req-8ea74433-628f-4d15-9d29-720aabfc940c service nova] Task: {'id': task-1293859, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.351768] env[63021]: DEBUG oslo_vmware.api [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293860, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.389911] env[63021]: DEBUG oslo_vmware.rw_handles [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5230102e-7773-2188-3adf-9cbf4ac916bf/disk-0.vmdk. {{(pid=63021) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 892.390960] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5e87b6e-f3e8-49d8-8ea8-bc0e3e594f02 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.398888] env[63021]: DEBUG oslo_vmware.rw_handles [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5230102e-7773-2188-3adf-9cbf4ac916bf/disk-0.vmdk is in state: ready. {{(pid=63021) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 892.399161] env[63021]: ERROR oslo_vmware.rw_handles [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5230102e-7773-2188-3adf-9cbf4ac916bf/disk-0.vmdk due to incomplete transfer. [ 892.399408] env[63021]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-d18ffdf8-ef60-4e70-85a8-2986f83cae22 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.409250] env[63021]: DEBUG oslo_vmware.rw_handles [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5230102e-7773-2188-3adf-9cbf4ac916bf/disk-0.vmdk. {{(pid=63021) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 892.409250] env[63021]: DEBUG nova.virt.vmwareapi.images [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Uploaded image 02de5fc5-60f6-49e2-ae03-e14c840dae1b to the Glance image server {{(pid=63021) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 892.410661] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Destroying the VM {{(pid=63021) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 892.410959] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-900700aa-cfe7-4c00-9b39-16bed900d984 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.426485] env[63021]: DEBUG oslo_vmware.api [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 892.426485] env[63021]: value = "task-1293863" [ 892.426485] env[63021]: _type = "Task" [ 892.426485] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.427670] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 892.427976] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Starting heal instance info cache {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 892.429048] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Rebuilding the list of instances to heal {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 892.440408] env[63021]: DEBUG oslo_vmware.api [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293863, 'name': Destroy_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.583195] env[63021]: DEBUG nova.scheduler.client.report [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 892.643428] env[63021]: DEBUG oslo_concurrency.lockutils [req-227023fa-159b-4e6b-9bd5-4071fb5a5e92 req-dc68025a-81f9-4788-940e-bca9bd14141f service nova] Releasing lock "refresh_cache-80b857c4-d9dd-4483-970f-a0f5f4bcf173" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.644087] env[63021]: DEBUG nova.compute.manager [req-227023fa-159b-4e6b-9bd5-4071fb5a5e92 req-dc68025a-81f9-4788-940e-bca9bd14141f service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Received event network-vif-deleted-d92edf1f-d15d-4f56-9da3-92cbe44b6709 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 892.644510] env[63021]: INFO nova.compute.manager [req-227023fa-159b-4e6b-9bd5-4071fb5a5e92 req-dc68025a-81f9-4788-940e-bca9bd14141f service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Neutron deleted interface d92edf1f-d15d-4f56-9da3-92cbe44b6709; detaching it from the instance and deleting it from the info cache [ 892.645292] env[63021]: DEBUG nova.network.neutron [req-227023fa-159b-4e6b-9bd5-4071fb5a5e92 req-dc68025a-81f9-4788-940e-bca9bd14141f service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Updating instance_info_cache with network_info: [{"id": "37dcb3f1-2931-40c6-96ba-b31b8fa1a72b", "address": "fa:16:3e:57:13:17", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37dcb3f1-29", "ovs_interfaceid": "37dcb3f1-2931-40c6-96ba-b31b8fa1a72b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2f88c221-8f47-478e-a5f3-ba0a230c7ad4", "address": "fa:16:3e:c2:01:70", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f88c221-8f", "ovs_interfaceid": "2f88c221-8f47-478e-a5f3-ba0a230c7ad4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.694708] env[63021]: DEBUG oslo_vmware.api [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293861, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.158704} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.697889] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 892.699251] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d022c22f-351c-45e8-9f37-e2d6c899eae7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.735444] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] 1b4871cc-3dac-4683-8fce-93c4d9e4407d/1b4871cc-3dac-4683-8fce-93c4d9e4407d.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 892.738947] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-478988bd-4681-44fb-8cf8-15af775b20b8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.756145] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dcf93f26-6ad3-471b-b85c-2f39720207f0 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Releasing lock "refresh_cache-36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.778497] env[63021]: DEBUG oslo_vmware.api [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 892.778497] env[63021]: value = "task-1293864" [ 892.778497] env[63021]: _type = "Task" [ 892.778497] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.784168] env[63021]: DEBUG oslo_vmware.api [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Task: {'id': task-1293862, 'name': RelocateVM_Task} progress is 38%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.796495] env[63021]: DEBUG oslo_vmware.api [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293864, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.820704] env[63021]: DEBUG oslo_vmware.api [req-492ee330-42de-456c-ae0f-7b6b5035ee63 req-8ea74433-628f-4d15-9d29-720aabfc940c service nova] Task: {'id': task-1293859, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.856521] env[63021]: DEBUG oslo_vmware.api [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293860, 'name': PowerOnVM_Task, 'duration_secs': 0.661396} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.856900] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 892.857191] env[63021]: INFO nova.compute.manager [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Took 9.14 seconds to spawn the instance on the hypervisor. [ 892.857534] env[63021]: DEBUG nova.compute.manager [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 892.858409] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f92a24a3-c1b9-46c1-99d8-112a92bca984 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.929201] env[63021]: DEBUG nova.network.neutron [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Updating instance_info_cache with network_info: [{"id": "1019be13-e855-4eab-8201-a2ae2b22476c", "address": "fa:16:3e:32:70:89", "network": {"id": "708d0dc3-c651-4001-9638-1d3538ff76c9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-983471056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "805af233efa147b3982fb794cd9e6f73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1019be13-e8", "ovs_interfaceid": "1019be13-e855-4eab-8201-a2ae2b22476c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.942025] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Skipping network cache update for instance because it is being deleted. {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 892.942025] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Skipping network cache update for instance because it is Building. {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 892.942025] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Skipping network cache update for instance because it is Building. {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 892.942025] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Skipping network cache update for instance because it is Building. {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 892.954174] env[63021]: DEBUG oslo_vmware.api [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293863, 'name': Destroy_Task} progress is 33%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.977760] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquiring lock "refresh_cache-7c6d7839-616d-41f5-a909-f8c626477a5e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.977760] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquired lock "refresh_cache-7c6d7839-616d-41f5-a909-f8c626477a5e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.977760] env[63021]: DEBUG nova.network.neutron [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Forcefully refreshing network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 892.977760] env[63021]: DEBUG nova.objects.instance [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lazy-loading 'info_cache' on Instance uuid 7c6d7839-616d-41f5-a909-f8c626477a5e {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 893.009029] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquiring lock "f0154308-e6a8-4321-a082-99291344664e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.009029] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Lock "f0154308-e6a8-4321-a082-99291344664e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.009029] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquiring lock "f0154308-e6a8-4321-a082-99291344664e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.009029] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Lock "f0154308-e6a8-4321-a082-99291344664e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.009029] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Lock "f0154308-e6a8-4321-a082-99291344664e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.011065] env[63021]: INFO nova.compute.manager [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Terminating instance [ 893.014089] env[63021]: DEBUG nova.compute.manager [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 893.014542] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 893.015709] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1e54a13-a605-47bf-98d0-5181c82f757e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.030864] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 893.031461] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-11efbe8d-7a6b-41e3-a4df-47a3afc08f3c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.042984] env[63021]: DEBUG oslo_vmware.api [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for the task: (returnval){ [ 893.042984] env[63021]: value = "task-1293865" [ 893.042984] env[63021]: _type = "Task" [ 893.042984] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.056810] env[63021]: DEBUG oslo_vmware.api [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293865, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.088746] env[63021]: DEBUG oslo_concurrency.lockutils [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.021s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.091620] env[63021]: DEBUG oslo_concurrency.lockutils [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.647s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.093600] env[63021]: INFO nova.compute.claims [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 893.111648] env[63021]: INFO nova.scheduler.client.report [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Deleted allocations for instance 6f4e6245-c0d9-46b4-baf5-5aca19e66da5 [ 893.151743] env[63021]: DEBUG oslo_concurrency.lockutils [req-227023fa-159b-4e6b-9bd5-4071fb5a5e92 req-dc68025a-81f9-4788-940e-bca9bd14141f service nova] Acquiring lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.272337] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dcf93f26-6ad3-471b-b85c-2f39720207f0 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "interface-36ec45ec-761b-4d62-a74f-e4d9a840ada0-d92edf1f-d15d-4f56-9da3-92cbe44b6709" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.684s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.274127] env[63021]: DEBUG oslo_vmware.api [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Task: {'id': task-1293862, 'name': RelocateVM_Task} progress is 53%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.299144] env[63021]: DEBUG oslo_vmware.api [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293864, 'name': ReconfigVM_Task, 'duration_secs': 0.480634} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.299499] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Reconfigured VM instance instance-0000004c to attach disk [datastore2] 1b4871cc-3dac-4683-8fce-93c4d9e4407d/1b4871cc-3dac-4683-8fce-93c4d9e4407d.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 893.300347] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b01af440-4afb-437d-84fd-b0648b8f126b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.313216] env[63021]: DEBUG oslo_vmware.api [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 893.313216] env[63021]: value = "task-1293866" [ 893.313216] env[63021]: _type = "Task" [ 893.313216] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.323820] env[63021]: DEBUG oslo_vmware.api [req-492ee330-42de-456c-ae0f-7b6b5035ee63 req-8ea74433-628f-4d15-9d29-720aabfc940c service nova] Task: {'id': task-1293859, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.333716] env[63021]: DEBUG oslo_vmware.api [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293866, 'name': Rename_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.385783] env[63021]: INFO nova.compute.manager [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Took 35.34 seconds to build instance. [ 893.436331] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Releasing lock "refresh_cache-71909213-258d-45f5-9a3e-e473e8ce8aa2" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.441222] env[63021]: DEBUG oslo_vmware.api [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293863, 'name': Destroy_Task, 'duration_secs': 0.834189} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.441836] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Destroyed the VM [ 893.442216] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Deleting Snapshot of the VM instance {{(pid=63021) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 893.442515] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-f8ecac46-b3c4-406d-820e-03b4908e4020 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.453160] env[63021]: DEBUG oslo_vmware.api [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 893.453160] env[63021]: value = "task-1293867" [ 893.453160] env[63021]: _type = "Task" [ 893.453160] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.467948] env[63021]: DEBUG oslo_vmware.api [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293867, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.558474] env[63021]: DEBUG oslo_vmware.api [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293865, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.621603] env[63021]: DEBUG oslo_concurrency.lockutils [None req-68959091-f868-41fa-90ea-702bde12543f tempest-ServerShowV257Test-320822938 tempest-ServerShowV257Test-320822938-project-member] Lock "6f4e6245-c0d9-46b4-baf5-5aca19e66da5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.233s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.684751] env[63021]: DEBUG oslo_concurrency.lockutils [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "d63fd465-975f-42ce-b1be-ac8a6929bc4d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.686809] env[63021]: DEBUG oslo_concurrency.lockutils [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "d63fd465-975f-42ce-b1be-ac8a6929bc4d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.769617] env[63021]: DEBUG oslo_vmware.api [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Task: {'id': task-1293862, 'name': RelocateVM_Task} progress is 65%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.778876] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2af3790-c86b-4f60-896d-29a9f39d263d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.788848] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c582a4d8-6b15-468e-bc96-a01756a8413b tempest-ServersAdminNegativeTestJSON-484360323 tempest-ServersAdminNegativeTestJSON-484360323-project-admin] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Suspending the VM {{(pid=63021) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 893.789250] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-e6e99d03-2dcc-4ed8-8c82-9adbc4cd5e7e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.800726] env[63021]: DEBUG oslo_vmware.api [None req-c582a4d8-6b15-468e-bc96-a01756a8413b tempest-ServersAdminNegativeTestJSON-484360323 tempest-ServersAdminNegativeTestJSON-484360323-project-admin] Waiting for the task: (returnval){ [ 893.800726] env[63021]: value = "task-1293868" [ 893.800726] env[63021]: _type = "Task" [ 893.800726] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.812882] env[63021]: DEBUG oslo_vmware.api [None req-c582a4d8-6b15-468e-bc96-a01756a8413b tempest-ServersAdminNegativeTestJSON-484360323 tempest-ServersAdminNegativeTestJSON-484360323-project-admin] Task: {'id': task-1293868, 'name': SuspendVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.822490] env[63021]: DEBUG oslo_vmware.api [req-492ee330-42de-456c-ae0f-7b6b5035ee63 req-8ea74433-628f-4d15-9d29-720aabfc940c service nova] Task: {'id': task-1293859, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.830150] env[63021]: DEBUG oslo_vmware.api [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293866, 'name': Rename_Task, 'duration_secs': 0.189714} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.830483] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 893.830832] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a2a9f243-f962-480c-bd99-9b20c7dc7340 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.841751] env[63021]: DEBUG oslo_vmware.api [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 893.841751] env[63021]: value = "task-1293869" [ 893.841751] env[63021]: _type = "Task" [ 893.841751] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.855950] env[63021]: DEBUG oslo_vmware.api [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293869, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.889936] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ec0df3c4-c6b2-493a-82f8-747f3bb6f50f tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Lock "51e4a14c-4792-43cd-9e99-4e06a8f6d6fb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.381s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.968402] env[63021]: DEBUG oslo_vmware.api [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293867, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.062201] env[63021]: DEBUG oslo_vmware.api [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293865, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.188863] env[63021]: DEBUG nova.compute.manager [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 894.273792] env[63021]: DEBUG oslo_vmware.api [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Task: {'id': task-1293862, 'name': RelocateVM_Task} progress is 81%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.317340] env[63021]: DEBUG oslo_vmware.api [None req-c582a4d8-6b15-468e-bc96-a01756a8413b tempest-ServersAdminNegativeTestJSON-484360323 tempest-ServersAdminNegativeTestJSON-484360323-project-admin] Task: {'id': task-1293868, 'name': SuspendVM_Task} progress is 83%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.325019] env[63021]: DEBUG oslo_vmware.api [req-492ee330-42de-456c-ae0f-7b6b5035ee63 req-8ea74433-628f-4d15-9d29-720aabfc940c service nova] Task: {'id': task-1293859, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.354921] env[63021]: DEBUG oslo_vmware.api [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293869, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.433637] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e31d289e-eb2d-46dd-bf95-5b313ce08ab0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.448028] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b86bfd9-7bd5-4aa4-9055-863b410dee8a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.495838] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1719493a-78a7-4aa3-be19-810754229a3a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.500163] env[63021]: DEBUG oslo_vmware.api [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293867, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.510724] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69dbd67f-9a0a-4cbb-81f4-266949274d84 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.528887] env[63021]: DEBUG nova.compute.provider_tree [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 894.561596] env[63021]: DEBUG oslo_vmware.api [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293865, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.709426] env[63021]: DEBUG oslo_concurrency.lockutils [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.744158] env[63021]: DEBUG nova.network.neutron [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Updating instance_info_cache with network_info: [{"id": "9ba9d35e-148e-45b3-a105-3ed457c3803c", "address": "fa:16:3e:6a:bd:ae", "network": {"id": "d67a9f2c-02da-411f-81dd-e2d702d6dd8b", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.35", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "cc50c08b679248798f76b07c1530ed7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "503991c4-44d0-42d9-aa03-5259331f1051", "external-id": "nsx-vlan-transportzone-3", "segmentation_id": 3, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9ba9d35e-14", "ovs_interfaceid": "9ba9d35e-148e-45b3-a105-3ed457c3803c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.767907] env[63021]: DEBUG oslo_vmware.api [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Task: {'id': task-1293862, 'name': RelocateVM_Task} progress is 95%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.811994] env[63021]: DEBUG oslo_vmware.api [None req-c582a4d8-6b15-468e-bc96-a01756a8413b tempest-ServersAdminNegativeTestJSON-484360323 tempest-ServersAdminNegativeTestJSON-484360323-project-admin] Task: {'id': task-1293868, 'name': SuspendVM_Task} progress is 83%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.820339] env[63021]: DEBUG oslo_vmware.api [req-492ee330-42de-456c-ae0f-7b6b5035ee63 req-8ea74433-628f-4d15-9d29-720aabfc940c service nova] Task: {'id': task-1293859, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.853937] env[63021]: DEBUG oslo_vmware.api [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293869, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.962638] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f0fc4af-b664-4481-8497-22dc0e45e415 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.985242] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Updating instance '71909213-258d-45f5-9a3e-e473e8ce8aa2' progress to 0 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 894.991426] env[63021]: DEBUG oslo_vmware.api [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293867, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.032688] env[63021]: DEBUG nova.scheduler.client.report [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 895.056039] env[63021]: DEBUG oslo_vmware.api [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293865, 'name': PowerOffVM_Task, 'duration_secs': 2.010591} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.056354] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 895.056523] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 895.056835] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1d5b17f7-e19d-4053-8c3c-e0e1e364dc8d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.143026] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 895.143026] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 895.143026] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Deleting the datastore file [datastore1] f0154308-e6a8-4321-a082-99291344664e {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 895.143026] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e17f6a02-4c2c-481a-9fb2-b7ea5471474d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.149198] env[63021]: DEBUG oslo_vmware.api [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for the task: (returnval){ [ 895.149198] env[63021]: value = "task-1293871" [ 895.149198] env[63021]: _type = "Task" [ 895.149198] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.158867] env[63021]: DEBUG oslo_vmware.api [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293871, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.247265] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Releasing lock "refresh_cache-7c6d7839-616d-41f5-a909-f8c626477a5e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.247507] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Updated the network info_cache for instance {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 895.247717] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 895.247966] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 895.248179] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 895.248364] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 895.248547] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 895.248729] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 895.248859] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63021) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 895.249011] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 895.268528] env[63021]: DEBUG oslo_vmware.api [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Task: {'id': task-1293862, 'name': RelocateVM_Task} progress is 97%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.311995] env[63021]: DEBUG oslo_vmware.api [None req-c582a4d8-6b15-468e-bc96-a01756a8413b tempest-ServersAdminNegativeTestJSON-484360323 tempest-ServersAdminNegativeTestJSON-484360323-project-admin] Task: {'id': task-1293868, 'name': SuspendVM_Task, 'duration_secs': 1.260827} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.312505] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c582a4d8-6b15-468e-bc96-a01756a8413b tempest-ServersAdminNegativeTestJSON-484360323 tempest-ServersAdminNegativeTestJSON-484360323-project-admin] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Suspended the VM {{(pid=63021) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 895.312728] env[63021]: DEBUG nova.compute.manager [None req-c582a4d8-6b15-468e-bc96-a01756a8413b tempest-ServersAdminNegativeTestJSON-484360323 tempest-ServersAdminNegativeTestJSON-484360323-project-admin] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 895.316375] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f4d9a48-5323-42dd-ae43-8b83d9ca3ab5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.324537] env[63021]: DEBUG oslo_vmware.api [req-492ee330-42de-456c-ae0f-7b6b5035ee63 req-8ea74433-628f-4d15-9d29-720aabfc940c service nova] Task: {'id': task-1293859, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.353061] env[63021]: DEBUG oslo_vmware.api [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293869, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.468769] env[63021]: DEBUG oslo_vmware.api [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293867, 'name': RemoveSnapshot_Task, 'duration_secs': 1.557152} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.469083] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Deleted Snapshot of the VM instance {{(pid=63021) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 895.469323] env[63021]: INFO nova.compute.manager [None req-0f616eb3-d427-407b-960e-61cb431d0244 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Took 17.75 seconds to snapshot the instance on the hypervisor. [ 895.494551] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 895.494699] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-78c8fed1-830d-4766-8231-b041df8c3569 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.502972] env[63021]: DEBUG oslo_vmware.api [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 895.502972] env[63021]: value = "task-1293872" [ 895.502972] env[63021]: _type = "Task" [ 895.502972] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.512513] env[63021]: DEBUG oslo_vmware.api [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1293872, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.536938] env[63021]: DEBUG oslo_concurrency.lockutils [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.445s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.537604] env[63021]: DEBUG nova.compute.manager [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 895.540277] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.800s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.541833] env[63021]: INFO nova.compute.claims [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 895.663030] env[63021]: DEBUG oslo_vmware.api [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293871, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.751954] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.769667] env[63021]: DEBUG oslo_vmware.api [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Task: {'id': task-1293862, 'name': RelocateVM_Task} progress is 98%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.822905] env[63021]: DEBUG oslo_vmware.api [req-492ee330-42de-456c-ae0f-7b6b5035ee63 req-8ea74433-628f-4d15-9d29-720aabfc940c service nova] Task: {'id': task-1293859, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.855728] env[63021]: DEBUG oslo_vmware.api [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293869, 'name': PowerOnVM_Task} progress is 78%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.014400] env[63021]: DEBUG oslo_vmware.api [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1293872, 'name': PowerOffVM_Task, 'duration_secs': 0.227691} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.014781] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 896.015086] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Updating instance '71909213-258d-45f5-9a3e-e473e8ce8aa2' progress to 17 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 896.046724] env[63021]: DEBUG nova.compute.utils [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 896.051442] env[63021]: DEBUG nova.compute.manager [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Not allocating networking since 'none' was specified. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 896.160589] env[63021]: DEBUG oslo_vmware.api [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293871, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.844761} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.160857] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 896.161078] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 896.161280] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 896.161502] env[63021]: INFO nova.compute.manager [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: f0154308-e6a8-4321-a082-99291344664e] Took 3.15 seconds to destroy the instance on the hypervisor. [ 896.161753] env[63021]: DEBUG oslo.service.loopingcall [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 896.161951] env[63021]: DEBUG nova.compute.manager [-] [instance: f0154308-e6a8-4321-a082-99291344664e] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 896.162063] env[63021]: DEBUG nova.network.neutron [-] [instance: f0154308-e6a8-4321-a082-99291344664e] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 896.270531] env[63021]: DEBUG oslo_vmware.api [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Task: {'id': task-1293862, 'name': RelocateVM_Task, 'duration_secs': 3.590203} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.270847] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Volume attach. Driver type: vmdk {{(pid=63021) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 896.271075] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277493', 'volume_id': 'e424adc5-a8b0-4409-b3e0-f3cdbc8fb712', 'name': 'volume-e424adc5-a8b0-4409-b3e0-f3cdbc8fb712', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '80b857c4-d9dd-4483-970f-a0f5f4bcf173', 'attached_at': '', 'detached_at': '', 'volume_id': 'e424adc5-a8b0-4409-b3e0-f3cdbc8fb712', 'serial': 'e424adc5-a8b0-4409-b3e0-f3cdbc8fb712'} {{(pid=63021) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 896.271866] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6d903c4-0cb5-4ca4-991c-e11caff4f7e8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.288838] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e9eae80-2ff2-410b-813f-9e1a54eae201 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.312289] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] volume-e424adc5-a8b0-4409-b3e0-f3cdbc8fb712/volume-e424adc5-a8b0-4409-b3e0-f3cdbc8fb712.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 896.312599] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cccf9fa4-82e7-4fb0-a4e5-46ea0c9b7e74 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.337186] env[63021]: DEBUG oslo_vmware.api [req-492ee330-42de-456c-ae0f-7b6b5035ee63 req-8ea74433-628f-4d15-9d29-720aabfc940c service nova] Task: {'id': task-1293859, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.338616] env[63021]: DEBUG oslo_vmware.api [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Waiting for the task: (returnval){ [ 896.338616] env[63021]: value = "task-1293873" [ 896.338616] env[63021]: _type = "Task" [ 896.338616] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.348006] env[63021]: DEBUG oslo_vmware.api [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Task: {'id': task-1293873, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.357070] env[63021]: DEBUG oslo_vmware.api [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293869, 'name': PowerOnVM_Task, 'duration_secs': 2.305781} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.357480] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 896.357753] env[63021]: INFO nova.compute.manager [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Took 10.09 seconds to spawn the instance on the hypervisor. [ 896.357997] env[63021]: DEBUG nova.compute.manager [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 896.358927] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ffb69eb-eeac-404c-82e6-ce328504cfce {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.453528] env[63021]: DEBUG nova.compute.manager [req-6175ef19-e399-468c-a997-0a19529049fd req-853b0049-bb7a-40c4-b9f8-c0d4b0a0d77a service nova] [instance: f0154308-e6a8-4321-a082-99291344664e] Received event network-vif-deleted-ce58eb3e-a5f1-4d1a-ab2a-dd6fcb0d4f62 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 896.453838] env[63021]: INFO nova.compute.manager [req-6175ef19-e399-468c-a997-0a19529049fd req-853b0049-bb7a-40c4-b9f8-c0d4b0a0d77a service nova] [instance: f0154308-e6a8-4321-a082-99291344664e] Neutron deleted interface ce58eb3e-a5f1-4d1a-ab2a-dd6fcb0d4f62; detaching it from the instance and deleting it from the info cache [ 896.455034] env[63021]: DEBUG nova.network.neutron [req-6175ef19-e399-468c-a997-0a19529049fd req-853b0049-bb7a-40c4-b9f8-c0d4b0a0d77a service nova] [instance: f0154308-e6a8-4321-a082-99291344664e] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.523206] env[63021]: DEBUG nova.virt.hardware [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 896.523517] env[63021]: DEBUG nova.virt.hardware [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 896.523745] env[63021]: DEBUG nova.virt.hardware [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 896.523996] env[63021]: DEBUG nova.virt.hardware [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 896.524214] env[63021]: DEBUG nova.virt.hardware [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 896.524413] env[63021]: DEBUG nova.virt.hardware [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 896.524698] env[63021]: DEBUG nova.virt.hardware [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 896.524929] env[63021]: DEBUG nova.virt.hardware [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 896.525204] env[63021]: DEBUG nova.virt.hardware [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 896.525368] env[63021]: DEBUG nova.virt.hardware [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 896.525548] env[63021]: DEBUG nova.virt.hardware [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 896.531263] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a7b5dffb-b9d5-44fc-bd7a-d7c795899214 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.548669] env[63021]: DEBUG oslo_vmware.api [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 896.548669] env[63021]: value = "task-1293874" [ 896.548669] env[63021]: _type = "Task" [ 896.548669] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.552810] env[63021]: DEBUG nova.compute.manager [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 896.564780] env[63021]: DEBUG oslo_vmware.api [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1293874, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.818729] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b103a5e-bc8d-4353-8d2b-f4b34b4b066c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.826984] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dae4503-a300-4c07-bd18-db7df9be9307 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.864501] env[63021]: DEBUG oslo_vmware.api [req-492ee330-42de-456c-ae0f-7b6b5035ee63 req-8ea74433-628f-4d15-9d29-720aabfc940c service nova] Task: {'id': task-1293859, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.865859] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bfd9b1b2-3da1-4969-9916-ce226c3b8037 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Acquiring lock "51e4a14c-4792-43cd-9e99-4e06a8f6d6fb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.866115] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bfd9b1b2-3da1-4969-9916-ce226c3b8037 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Lock "51e4a14c-4792-43cd-9e99-4e06a8f6d6fb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.866319] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bfd9b1b2-3da1-4969-9916-ce226c3b8037 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Acquiring lock "51e4a14c-4792-43cd-9e99-4e06a8f6d6fb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.866505] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bfd9b1b2-3da1-4969-9916-ce226c3b8037 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Lock "51e4a14c-4792-43cd-9e99-4e06a8f6d6fb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.866674] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bfd9b1b2-3da1-4969-9916-ce226c3b8037 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Lock "51e4a14c-4792-43cd-9e99-4e06a8f6d6fb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.871670] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbbb87f7-9c46-4e0e-ab60-437a5e84001f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.876757] env[63021]: INFO nova.compute.manager [None req-bfd9b1b2-3da1-4969-9916-ce226c3b8037 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Terminating instance [ 896.885020] env[63021]: DEBUG nova.compute.manager [None req-bfd9b1b2-3da1-4969-9916-ce226c3b8037 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 896.885020] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-bfd9b1b2-3da1-4969-9916-ce226c3b8037 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 896.885020] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbc703d7-a15f-4902-b76d-f364c889937d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.893750] env[63021]: DEBUG oslo_vmware.api [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Task: {'id': task-1293873, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.894498] env[63021]: INFO nova.compute.manager [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Took 37.83 seconds to build instance. [ 896.896730] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8453951-60bf-4a86-9086-cd01574b7417 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.903440] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-bfd9b1b2-3da1-4969-9916-ce226c3b8037 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 896.904064] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-124a088e-acf8-4c08-85fb-40e45e6cb487 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.914796] env[63021]: DEBUG nova.compute.provider_tree [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 896.927021] env[63021]: DEBUG nova.network.neutron [-] [instance: f0154308-e6a8-4321-a082-99291344664e] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.957537] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d0b4adf7-d828-4325-9be4-137e1127187a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.970120] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed8a1bce-8e98-4559-bfa4-d2b0f786f258 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.001020] env[63021]: DEBUG nova.compute.manager [req-6175ef19-e399-468c-a997-0a19529049fd req-853b0049-bb7a-40c4-b9f8-c0d4b0a0d77a service nova] [instance: f0154308-e6a8-4321-a082-99291344664e] Detach interface failed, port_id=ce58eb3e-a5f1-4d1a-ab2a-dd6fcb0d4f62, reason: Instance f0154308-e6a8-4321-a082-99291344664e could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 897.063324] env[63021]: DEBUG oslo_vmware.api [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1293874, 'name': ReconfigVM_Task, 'duration_secs': 0.244142} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.063897] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Updating instance '71909213-258d-45f5-9a3e-e473e8ce8aa2' progress to 33 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 897.197870] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-bfd9b1b2-3da1-4969-9916-ce226c3b8037 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 897.198156] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-bfd9b1b2-3da1-4969-9916-ce226c3b8037 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Deleting contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 897.198438] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-bfd9b1b2-3da1-4969-9916-ce226c3b8037 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Deleting the datastore file [datastore2] 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 897.198771] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-371d82c1-f601-4bf2-a3e6-162906ef691b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.206662] env[63021]: DEBUG oslo_vmware.api [None req-bfd9b1b2-3da1-4969-9916-ce226c3b8037 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Waiting for the task: (returnval){ [ 897.206662] env[63021]: value = "task-1293876" [ 897.206662] env[63021]: _type = "Task" [ 897.206662] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.218962] env[63021]: DEBUG oslo_vmware.api [None req-bfd9b1b2-3da1-4969-9916-ce226c3b8037 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293876, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.340852] env[63021]: DEBUG oslo_vmware.api [req-492ee330-42de-456c-ae0f-7b6b5035ee63 req-8ea74433-628f-4d15-9d29-720aabfc940c service nova] Task: {'id': task-1293859, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.378538] env[63021]: DEBUG oslo_vmware.api [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Task: {'id': task-1293873, 'name': ReconfigVM_Task, 'duration_secs': 0.585714} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.378978] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Reconfigured VM instance instance-0000004d to attach disk [datastore2] volume-e424adc5-a8b0-4409-b3e0-f3cdbc8fb712/volume-e424adc5-a8b0-4409-b3e0-f3cdbc8fb712.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 897.384010] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-be7138ee-2ec0-402e-8db9-2d8a2b0123cc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.402173] env[63021]: DEBUG oslo_vmware.api [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Waiting for the task: (returnval){ [ 897.402173] env[63021]: value = "task-1293877" [ 897.402173] env[63021]: _type = "Task" [ 897.402173] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.405905] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2a857dae-c9e0-4164-810d-2056e6e2ddf8 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Lock "1b4871cc-3dac-4683-8fce-93c4d9e4407d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 58.261s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.413015] env[63021]: DEBUG oslo_vmware.api [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Task: {'id': task-1293877, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.418419] env[63021]: DEBUG nova.scheduler.client.report [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 897.430227] env[63021]: INFO nova.compute.manager [-] [instance: f0154308-e6a8-4321-a082-99291344664e] Took 1.27 seconds to deallocate network for instance. [ 897.565040] env[63021]: DEBUG nova.compute.manager [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 897.571726] env[63021]: DEBUG nova.virt.hardware [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 897.572083] env[63021]: DEBUG nova.virt.hardware [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 897.572309] env[63021]: DEBUG nova.virt.hardware [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 897.572606] env[63021]: DEBUG nova.virt.hardware [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 897.572837] env[63021]: DEBUG nova.virt.hardware [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 897.573044] env[63021]: DEBUG nova.virt.hardware [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 897.573338] env[63021]: DEBUG nova.virt.hardware [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 897.573563] env[63021]: DEBUG nova.virt.hardware [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 897.573809] env[63021]: DEBUG nova.virt.hardware [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 897.574055] env[63021]: DEBUG nova.virt.hardware [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 897.574299] env[63021]: DEBUG nova.virt.hardware [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 897.582773] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Reconfiguring VM instance instance-00000042 to detach disk 2000 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 897.583666] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fd26dee0-4c4d-4f2a-a745-b6e148bfcdc8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.615310] env[63021]: DEBUG nova.virt.hardware [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 897.615665] env[63021]: DEBUG nova.virt.hardware [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 897.615909] env[63021]: DEBUG nova.virt.hardware [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 897.616243] env[63021]: DEBUG nova.virt.hardware [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 897.616490] env[63021]: DEBUG nova.virt.hardware [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 897.616730] env[63021]: DEBUG nova.virt.hardware [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 897.617064] env[63021]: DEBUG nova.virt.hardware [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 897.617337] env[63021]: DEBUG nova.virt.hardware [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 897.617665] env[63021]: DEBUG nova.virt.hardware [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 897.617970] env[63021]: DEBUG nova.virt.hardware [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 897.618281] env[63021]: DEBUG nova.virt.hardware [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 897.619462] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b59e857-17a2-4d67-a149-d7bfe30d66b2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.625337] env[63021]: DEBUG oslo_vmware.api [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 897.625337] env[63021]: value = "task-1293878" [ 897.625337] env[63021]: _type = "Task" [ 897.625337] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.635879] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3006c5a1-8242-4fe8-a89b-07120082beda {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.647199] env[63021]: DEBUG oslo_vmware.api [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1293878, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.663015] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Instance VIF info [] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 897.672175] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Creating folder: Project (f9b299b877ab45f8941cf289d50e08e2). Parent ref: group-v277447. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 897.672609] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fe09b604-ff8d-408b-93e5-84ed80b06cf3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.688422] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Created folder: Project (f9b299b877ab45f8941cf289d50e08e2) in parent group-v277447. [ 897.688621] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Creating folder: Instances. Parent ref: group-v277536. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 897.688891] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b9346a92-9c0f-4ccb-86ab-38ad171bfee7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.699518] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Created folder: Instances in parent group-v277536. [ 897.699864] env[63021]: DEBUG oslo.service.loopingcall [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 897.700128] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 897.700392] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b3c37b69-4e7e-4387-ab5c-7fc53563628b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.721984] env[63021]: DEBUG oslo_vmware.api [None req-bfd9b1b2-3da1-4969-9916-ce226c3b8037 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293876, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.324282} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.723222] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-bfd9b1b2-3da1-4969-9916-ce226c3b8037 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 897.723441] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-bfd9b1b2-3da1-4969-9916-ce226c3b8037 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Deleted contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 897.723618] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-bfd9b1b2-3da1-4969-9916-ce226c3b8037 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 897.723793] env[63021]: INFO nova.compute.manager [None req-bfd9b1b2-3da1-4969-9916-ce226c3b8037 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Took 0.84 seconds to destroy the instance on the hypervisor. [ 897.724044] env[63021]: DEBUG oslo.service.loopingcall [None req-bfd9b1b2-3da1-4969-9916-ce226c3b8037 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 897.724234] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 897.724234] env[63021]: value = "task-1293881" [ 897.724234] env[63021]: _type = "Task" [ 897.724234] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.724416] env[63021]: DEBUG nova.compute.manager [-] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 897.724512] env[63021]: DEBUG nova.network.neutron [-] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 897.734997] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293881, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.842287] env[63021]: DEBUG oslo_vmware.api [req-492ee330-42de-456c-ae0f-7b6b5035ee63 req-8ea74433-628f-4d15-9d29-720aabfc940c service nova] Task: {'id': task-1293859, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.913215] env[63021]: DEBUG oslo_vmware.api [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Task: {'id': task-1293877, 'name': ReconfigVM_Task, 'duration_secs': 0.191688} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.913495] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277493', 'volume_id': 'e424adc5-a8b0-4409-b3e0-f3cdbc8fb712', 'name': 'volume-e424adc5-a8b0-4409-b3e0-f3cdbc8fb712', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '80b857c4-d9dd-4483-970f-a0f5f4bcf173', 'attached_at': '', 'detached_at': '', 'volume_id': 'e424adc5-a8b0-4409-b3e0-f3cdbc8fb712', 'serial': 'e424adc5-a8b0-4409-b3e0-f3cdbc8fb712'} {{(pid=63021) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 897.914389] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-af850649-14ff-4a75-a1e4-02b28a56fc3e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.924516] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.384s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.925060] env[63021]: DEBUG nova.compute.manager [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 897.927865] env[63021]: DEBUG oslo_vmware.api [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Waiting for the task: (returnval){ [ 897.927865] env[63021]: value = "task-1293882" [ 897.927865] env[63021]: _type = "Task" [ 897.927865] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.928443] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.061s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.928540] env[63021]: DEBUG nova.objects.instance [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Lazy-loading 'resources' on Instance uuid 7787f6bb-e855-476f-9146-2a648932429d {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 897.937318] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.941640] env[63021]: DEBUG oslo_vmware.api [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Task: {'id': task-1293882, 'name': Rename_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.052201] env[63021]: INFO nova.compute.manager [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Rescuing [ 898.052327] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquiring lock "refresh_cache-1b4871cc-3dac-4683-8fce-93c4d9e4407d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.052385] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquired lock "refresh_cache-1b4871cc-3dac-4683-8fce-93c4d9e4407d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.052555] env[63021]: DEBUG nova.network.neutron [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 898.137059] env[63021]: DEBUG oslo_vmware.api [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1293878, 'name': ReconfigVM_Task, 'duration_secs': 0.192217} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.137379] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Reconfigured VM instance instance-00000042 to detach disk 2000 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 898.138209] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb7576c7-b1eb-4575-b06e-6d509d71a492 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.162205] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] 71909213-258d-45f5-9a3e-e473e8ce8aa2/71909213-258d-45f5-9a3e-e473e8ce8aa2.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 898.162535] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd12d551-0293-4897-95c9-de4c470c9acf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.181275] env[63021]: DEBUG oslo_vmware.api [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 898.181275] env[63021]: value = "task-1293883" [ 898.181275] env[63021]: _type = "Task" [ 898.181275] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.190199] env[63021]: DEBUG oslo_vmware.api [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1293883, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.237548] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293881, 'name': CreateVM_Task, 'duration_secs': 0.481361} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.237714] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 898.238190] env[63021]: DEBUG oslo_concurrency.lockutils [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.238379] env[63021]: DEBUG oslo_concurrency.lockutils [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.239015] env[63021]: DEBUG oslo_concurrency.lockutils [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 898.239306] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1693688b-aa3d-45e7-9d1d-47351df95b7f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.245017] env[63021]: DEBUG oslo_vmware.api [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Waiting for the task: (returnval){ [ 898.245017] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]522e1efe-19d5-a52d-5f77-9e7009011b00" [ 898.245017] env[63021]: _type = "Task" [ 898.245017] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.254525] env[63021]: DEBUG oslo_vmware.api [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]522e1efe-19d5-a52d-5f77-9e7009011b00, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.343052] env[63021]: DEBUG oslo_vmware.api [req-492ee330-42de-456c-ae0f-7b6b5035ee63 req-8ea74433-628f-4d15-9d29-720aabfc940c service nova] Task: {'id': task-1293859, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.384353] env[63021]: DEBUG oslo_concurrency.lockutils [None req-22810790-0d2c-4212-8068-81258f676e07 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquiring lock "df856146-61ba-4de7-a5fd-d75263927163" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.384650] env[63021]: DEBUG oslo_concurrency.lockutils [None req-22810790-0d2c-4212-8068-81258f676e07 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "df856146-61ba-4de7-a5fd-d75263927163" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.430238] env[63021]: DEBUG nova.compute.utils [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 898.435117] env[63021]: DEBUG nova.compute.manager [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 898.435117] env[63021]: DEBUG nova.network.neutron [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 898.449756] env[63021]: DEBUG oslo_vmware.api [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Task: {'id': task-1293882, 'name': Rename_Task, 'duration_secs': 0.191278} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.450610] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 898.451829] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2379dd04-c421-4fab-9ab4-645fabd6f8f4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.465727] env[63021]: DEBUG oslo_vmware.api [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Waiting for the task: (returnval){ [ 898.465727] env[63021]: value = "task-1293884" [ 898.465727] env[63021]: _type = "Task" [ 898.465727] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.484151] env[63021]: DEBUG oslo_vmware.api [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Task: {'id': task-1293884, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.486317] env[63021]: DEBUG nova.policy [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ea10d6ee13544020a76db4ffec5ec41e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c9a23f10cf5a4b178c7efff86041ce21', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 898.488242] env[63021]: DEBUG nova.network.neutron [-] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.586712] env[63021]: DEBUG nova.compute.manager [req-58a612f0-c858-49d5-80c8-a68832844cb0 req-8f8cabbc-523f-45ee-b2b3-041d58cc20f2 service nova] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Received event network-vif-deleted-664021c6-82fb-46ea-a7d2-c9e942e0dded {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 898.696797] env[63021]: DEBUG oslo_vmware.api [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1293883, 'name': ReconfigVM_Task, 'duration_secs': 0.294761} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.697207] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Reconfigured VM instance instance-00000042 to attach disk [datastore2] 71909213-258d-45f5-9a3e-e473e8ce8aa2/71909213-258d-45f5-9a3e-e473e8ce8aa2.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 898.697459] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Updating instance '71909213-258d-45f5-9a3e-e473e8ce8aa2' progress to 50 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 898.751505] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b29a788-eca4-4cfa-b9b1-56b11706c4fb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.767849] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-508a53ce-ef1a-4163-beb5-acf250f24b8b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.771743] env[63021]: DEBUG oslo_vmware.api [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]522e1efe-19d5-a52d-5f77-9e7009011b00, 'name': SearchDatastore_Task, 'duration_secs': 0.013525} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.772458] env[63021]: DEBUG oslo_concurrency.lockutils [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.772725] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 898.772988] env[63021]: DEBUG oslo_concurrency.lockutils [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.773152] env[63021]: DEBUG oslo_concurrency.lockutils [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.773337] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 898.773999] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-13435d78-58b7-4b26-88db-b0c9438b8f5d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.803200] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d420a915-76e7-4837-8d72-1a5b7ed9119d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.805878] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 898.806091] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 898.806793] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e851e563-8e09-4d3d-8a17-1ff40cfef3cc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.812978] env[63021]: DEBUG oslo_vmware.api [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Waiting for the task: (returnval){ [ 898.812978] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]523ce537-13bc-3b58-ceab-6667f6444894" [ 898.812978] env[63021]: _type = "Task" [ 898.812978] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.819109] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-463dff22-2751-463e-b88b-72fdc7ab4153 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.830767] env[63021]: DEBUG oslo_vmware.api [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]523ce537-13bc-3b58-ceab-6667f6444894, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.840659] env[63021]: DEBUG nova.compute.provider_tree [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 898.850859] env[63021]: DEBUG oslo_vmware.api [req-492ee330-42de-456c-ae0f-7b6b5035ee63 req-8ea74433-628f-4d15-9d29-720aabfc940c service nova] Task: {'id': task-1293859, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.881979] env[63021]: DEBUG nova.network.neutron [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Updating instance_info_cache with network_info: [{"id": "d34cfd0d-464e-400b-8366-83321cc6112f", "address": "fa:16:3e:64:67:d0", "network": {"id": "37008b1f-68ff-404c-9afc-14b108af85c3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1473182647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "2d364c9eae4b4f8a8b60d4a3ac2ed648", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "02bbcead-d833-4543-bec6-fb82dfe659ff", "external-id": "nsx-vlan-transportzone-478", "segmentation_id": 478, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd34cfd0d-46", "ovs_interfaceid": "d34cfd0d-464e-400b-8366-83321cc6112f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.888072] env[63021]: DEBUG nova.compute.utils [None req-22810790-0d2c-4212-8068-81258f676e07 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 898.927122] env[63021]: DEBUG nova.network.neutron [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Successfully created port: 2e8a7fc6-825c-45a5-b41c-d28410061aa3 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 898.939775] env[63021]: DEBUG nova.compute.manager [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 898.980941] env[63021]: DEBUG oslo_vmware.api [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Task: {'id': task-1293884, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.990440] env[63021]: INFO nova.compute.manager [-] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Took 1.27 seconds to deallocate network for instance. [ 899.206280] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8251f263-204d-41d4-a6fe-eecda83b6b70 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.226469] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-991adcfe-cf41-441d-b5bf-6244c8c885a0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.245754] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Updating instance '71909213-258d-45f5-9a3e-e473e8ce8aa2' progress to 67 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 899.325783] env[63021]: DEBUG oslo_vmware.api [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]523ce537-13bc-3b58-ceab-6667f6444894, 'name': SearchDatastore_Task, 'duration_secs': 0.03051} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.326620] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8092c5d3-17c6-406a-b643-fd25ef21a9c1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.333137] env[63021]: DEBUG oslo_vmware.api [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Waiting for the task: (returnval){ [ 899.333137] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52469aca-e7a8-9f22-bfe4-0ad46099c594" [ 899.333137] env[63021]: _type = "Task" [ 899.333137] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.341183] env[63021]: DEBUG oslo_vmware.api [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52469aca-e7a8-9f22-bfe4-0ad46099c594, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.346079] env[63021]: DEBUG nova.scheduler.client.report [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 899.352231] env[63021]: DEBUG oslo_vmware.api [req-492ee330-42de-456c-ae0f-7b6b5035ee63 req-8ea74433-628f-4d15-9d29-720aabfc940c service nova] Task: {'id': task-1293859, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.385471] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Releasing lock "refresh_cache-1b4871cc-3dac-4683-8fce-93c4d9e4407d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.390933] env[63021]: DEBUG oslo_concurrency.lockutils [None req-22810790-0d2c-4212-8068-81258f676e07 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "df856146-61ba-4de7-a5fd-d75263927163" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.479062] env[63021]: DEBUG oslo_vmware.api [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Task: {'id': task-1293884, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.497880] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bfd9b1b2-3da1-4969-9916-ce226c3b8037 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.791598] env[63021]: DEBUG nova.network.neutron [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Port 1019be13-e855-4eab-8201-a2ae2b22476c binding to destination host cpu-1 is already ACTIVE {{(pid=63021) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 899.844130] env[63021]: DEBUG oslo_vmware.api [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52469aca-e7a8-9f22-bfe4-0ad46099c594, 'name': SearchDatastore_Task, 'duration_secs': 0.032494} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.849021] env[63021]: DEBUG oslo_concurrency.lockutils [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.849021] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 4d712cc1-6202-407e-8767-c620a1ce0bba/4d712cc1-6202-407e-8767-c620a1ce0bba.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 899.849021] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9235b982-d4c1-4a0a-9951-8fa0f6bbb272 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.854698] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.926s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.861308] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.612s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.861717] env[63021]: DEBUG nova.objects.instance [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lazy-loading 'resources' on Instance uuid 16b21798-17e8-4aeb-affa-57eae31d5dd3 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 899.863301] env[63021]: DEBUG oslo_vmware.api [req-492ee330-42de-456c-ae0f-7b6b5035ee63 req-8ea74433-628f-4d15-9d29-720aabfc940c service nova] Task: {'id': task-1293859, 'name': ReconfigVM_Task, 'duration_secs': 8.468506} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.863589] env[63021]: DEBUG oslo_vmware.api [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Waiting for the task: (returnval){ [ 899.863589] env[63021]: value = "task-1293885" [ 899.863589] env[63021]: _type = "Task" [ 899.863589] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.864018] env[63021]: DEBUG oslo_concurrency.lockutils [req-492ee330-42de-456c-ae0f-7b6b5035ee63 req-8ea74433-628f-4d15-9d29-720aabfc940c service nova] Releasing lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.864232] env[63021]: DEBUG nova.virt.vmwareapi.vmops [req-492ee330-42de-456c-ae0f-7b6b5035ee63 req-8ea74433-628f-4d15-9d29-720aabfc940c service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Reconfigured VM to detach interface {{(pid=63021) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 899.864880] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 8.592s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.865104] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.865360] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.865540] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.870611] env[63021]: INFO nova.compute.manager [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Terminating instance [ 899.872973] env[63021]: DEBUG nova.compute.manager [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 899.873202] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 899.874569] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39b7f35f-1f17-4ef6-a725-35edc994ece5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.881377] env[63021]: DEBUG oslo_vmware.api [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293885, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.884794] env[63021]: INFO nova.scheduler.client.report [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Deleted allocations for instance 7787f6bb-e855-476f-9146-2a648932429d [ 899.890019] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 899.891166] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7e778fb9-fe30-4d17-8a24-5d67d5a0b532 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.902677] env[63021]: DEBUG oslo_vmware.api [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 899.902677] env[63021]: value = "task-1293886" [ 899.902677] env[63021]: _type = "Task" [ 899.902677] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.919715] env[63021]: DEBUG oslo_vmware.api [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293886, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.920980] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 899.921354] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c14bf37c-789a-4e65-8e01-c2761b6c48fb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.932410] env[63021]: DEBUG oslo_vmware.api [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 899.932410] env[63021]: value = "task-1293887" [ 899.932410] env[63021]: _type = "Task" [ 899.932410] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.943050] env[63021]: DEBUG oslo_vmware.api [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293887, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.951301] env[63021]: DEBUG nova.compute.manager [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 899.974191] env[63021]: DEBUG nova.virt.hardware [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 899.974512] env[63021]: DEBUG nova.virt.hardware [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 899.974930] env[63021]: DEBUG nova.virt.hardware [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 899.975584] env[63021]: DEBUG nova.virt.hardware [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 899.975584] env[63021]: DEBUG nova.virt.hardware [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 899.975584] env[63021]: DEBUG nova.virt.hardware [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 899.975781] env[63021]: DEBUG nova.virt.hardware [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 899.975994] env[63021]: DEBUG nova.virt.hardware [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 899.976166] env[63021]: DEBUG nova.virt.hardware [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 899.976352] env[63021]: DEBUG nova.virt.hardware [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 899.976531] env[63021]: DEBUG nova.virt.hardware [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 899.977580] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f4152ca-3299-484d-a2cf-8815e8dab892 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.988762] env[63021]: DEBUG oslo_vmware.api [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Task: {'id': task-1293884, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.995938] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b043b208-3c2e-424f-8200-03dcec5024c6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.382145] env[63021]: DEBUG oslo_vmware.api [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293885, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.399898] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c9853b92-6a6d-4377-954c-7cb2493b6bae tempest-ServerRescueTestJSONUnderV235-1720159489 tempest-ServerRescueTestJSONUnderV235-1720159489-project-member] Lock "7787f6bb-e855-476f-9146-2a648932429d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.315s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.416012] env[63021]: DEBUG oslo_vmware.api [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293886, 'name': PowerOffVM_Task, 'duration_secs': 0.22153} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.419618] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 900.419797] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 900.420823] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a516b706-ebae-4add-885b-aa3b72fb8da7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.448376] env[63021]: DEBUG oslo_vmware.api [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293887, 'name': PowerOffVM_Task, 'duration_secs': 0.191742} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.448376] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 900.449303] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1e7545d-5438-4c1c-8c8e-faa49bd05390 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.479349] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a589fecd-848e-42bc-af82-ba26e53911bf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.486111] env[63021]: DEBUG nova.network.neutron [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Successfully updated port: 2e8a7fc6-825c-45a5-b41c-d28410061aa3 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 900.491028] env[63021]: DEBUG oslo_concurrency.lockutils [None req-22810790-0d2c-4212-8068-81258f676e07 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquiring lock "df856146-61ba-4de7-a5fd-d75263927163" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.491148] env[63021]: DEBUG oslo_concurrency.lockutils [None req-22810790-0d2c-4212-8068-81258f676e07 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "df856146-61ba-4de7-a5fd-d75263927163" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.491566] env[63021]: INFO nova.compute.manager [None req-22810790-0d2c-4212-8068-81258f676e07 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Attaching volume 1d748e34-0bec-4d79-9a3a-f42d78520ff7 to /dev/sdb [ 900.509589] env[63021]: DEBUG oslo_vmware.api [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Task: {'id': task-1293884, 'name': PowerOnVM_Task, 'duration_secs': 1.871516} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.512964] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 900.512964] env[63021]: INFO nova.compute.manager [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Took 10.09 seconds to spawn the instance on the hypervisor. [ 900.512964] env[63021]: DEBUG nova.compute.manager [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 900.516099] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 900.516099] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 900.516099] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Deleting the datastore file [datastore1] 36ec45ec-761b-4d62-a74f-e4d9a840ada0 {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 900.516099] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d92a5bc8-bdb5-454e-89b5-d5b68cb65e3e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.517888] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-92a7f6c8-a320-4241-8c58-85c1a307e945 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.537403] env[63021]: DEBUG oslo_vmware.api [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 900.537403] env[63021]: value = "task-1293889" [ 900.537403] env[63021]: _type = "Task" [ 900.537403] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.539436] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 900.539760] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8f8b99b7-2f07-46ec-92ac-24dead26cefd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.558310] env[63021]: DEBUG oslo_vmware.api [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293889, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.558310] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0107eafe-af74-4a4f-8374-d0a3c3b8d85e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.559960] env[63021]: DEBUG oslo_vmware.api [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 900.559960] env[63021]: value = "task-1293890" [ 900.559960] env[63021]: _type = "Task" [ 900.559960] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.570560] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6be43abe-2b16-48a5-b184-bfd30813b745 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.576867] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] VM already powered off {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 900.577147] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 900.577417] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.577570] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.577746] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 900.578278] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d19101f3-c32f-4a16-88e9-431700f5dae2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.590448] env[63021]: DEBUG nova.virt.block_device [None req-22810790-0d2c-4212-8068-81258f676e07 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Updating existing volume attachment record: 642fd66b-8473-4219-85d9-10a41fd9af13 {{(pid=63021) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 900.595965] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 900.596164] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 900.597169] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a24ea9c4-ec10-43c3-b2f9-8e544844ad4b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.605213] env[63021]: DEBUG oslo_vmware.api [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 900.605213] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]524d1ad1-826a-981b-aae7-db97e2ebf9f3" [ 900.605213] env[63021]: _type = "Task" [ 900.605213] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.619622] env[63021]: DEBUG oslo_vmware.api [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]524d1ad1-826a-981b-aae7-db97e2ebf9f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.642611] env[63021]: DEBUG nova.compute.manager [req-44951688-29d9-4e2c-b3f2-c3bf334ebedc req-8589408c-ca9a-4d7f-846a-d8b92ea8d8cb service nova] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Received event network-vif-plugged-2e8a7fc6-825c-45a5-b41c-d28410061aa3 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 900.643199] env[63021]: DEBUG oslo_concurrency.lockutils [req-44951688-29d9-4e2c-b3f2-c3bf334ebedc req-8589408c-ca9a-4d7f-846a-d8b92ea8d8cb service nova] Acquiring lock "ed26dff1-fed0-4baf-ad41-d14850254aec-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.643331] env[63021]: DEBUG oslo_concurrency.lockutils [req-44951688-29d9-4e2c-b3f2-c3bf334ebedc req-8589408c-ca9a-4d7f-846a-d8b92ea8d8cb service nova] Lock "ed26dff1-fed0-4baf-ad41-d14850254aec-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.643494] env[63021]: DEBUG oslo_concurrency.lockutils [req-44951688-29d9-4e2c-b3f2-c3bf334ebedc req-8589408c-ca9a-4d7f-846a-d8b92ea8d8cb service nova] Lock "ed26dff1-fed0-4baf-ad41-d14850254aec-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.643670] env[63021]: DEBUG nova.compute.manager [req-44951688-29d9-4e2c-b3f2-c3bf334ebedc req-8589408c-ca9a-4d7f-846a-d8b92ea8d8cb service nova] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] No waiting events found dispatching network-vif-plugged-2e8a7fc6-825c-45a5-b41c-d28410061aa3 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 900.643829] env[63021]: WARNING nova.compute.manager [req-44951688-29d9-4e2c-b3f2-c3bf334ebedc req-8589408c-ca9a-4d7f-846a-d8b92ea8d8cb service nova] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Received unexpected event network-vif-plugged-2e8a7fc6-825c-45a5-b41c-d28410061aa3 for instance with vm_state building and task_state spawning. [ 900.644077] env[63021]: DEBUG nova.compute.manager [req-44951688-29d9-4e2c-b3f2-c3bf334ebedc req-8589408c-ca9a-4d7f-846a-d8b92ea8d8cb service nova] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Received event network-changed-2e8a7fc6-825c-45a5-b41c-d28410061aa3 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 900.644326] env[63021]: DEBUG nova.compute.manager [req-44951688-29d9-4e2c-b3f2-c3bf334ebedc req-8589408c-ca9a-4d7f-846a-d8b92ea8d8cb service nova] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Refreshing instance network info cache due to event network-changed-2e8a7fc6-825c-45a5-b41c-d28410061aa3. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 900.644517] env[63021]: DEBUG oslo_concurrency.lockutils [req-44951688-29d9-4e2c-b3f2-c3bf334ebedc req-8589408c-ca9a-4d7f-846a-d8b92ea8d8cb service nova] Acquiring lock "refresh_cache-ed26dff1-fed0-4baf-ad41-d14850254aec" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.644654] env[63021]: DEBUG oslo_concurrency.lockutils [req-44951688-29d9-4e2c-b3f2-c3bf334ebedc req-8589408c-ca9a-4d7f-846a-d8b92ea8d8cb service nova] Acquired lock "refresh_cache-ed26dff1-fed0-4baf-ad41-d14850254aec" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.644812] env[63021]: DEBUG nova.network.neutron [req-44951688-29d9-4e2c-b3f2-c3bf334ebedc req-8589408c-ca9a-4d7f-846a-d8b92ea8d8cb service nova] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Refreshing network info cache for port 2e8a7fc6-825c-45a5-b41c-d28410061aa3 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 900.712563] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1e84d5c-f669-4f79-89b0-dc50829085bd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.721183] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-788a6aab-1ea6-41fd-b702-c13ff802a985 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.752684] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-965170d8-5d46-4774-b0c2-56d90863961e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.760599] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e60cfdf-6065-4a2c-8487-5fe2b106e226 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.774895] env[63021]: DEBUG nova.compute.provider_tree [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 900.815324] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "71909213-258d-45f5-9a3e-e473e8ce8aa2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.815627] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "71909213-258d-45f5-9a3e-e473e8ce8aa2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.815842] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "71909213-258d-45f5-9a3e-e473e8ce8aa2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.878635] env[63021]: DEBUG oslo_vmware.api [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293885, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.652039} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.878927] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 4d712cc1-6202-407e-8767-c620a1ce0bba/4d712cc1-6202-407e-8767-c620a1ce0bba.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 900.879188] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 900.879501] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-56777733-a515-4f63-8080-af0a00ac8cff {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.887173] env[63021]: DEBUG oslo_vmware.api [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Waiting for the task: (returnval){ [ 900.887173] env[63021]: value = "task-1293892" [ 900.887173] env[63021]: _type = "Task" [ 900.887173] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.898921] env[63021]: DEBUG oslo_vmware.api [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293892, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.989421] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Acquiring lock "refresh_cache-ed26dff1-fed0-4baf-ad41-d14850254aec" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.050924] env[63021]: INFO nova.compute.manager [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Took 40.31 seconds to build instance. [ 901.060788] env[63021]: DEBUG oslo_vmware.api [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293889, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.49889} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.063145] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 901.063366] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 901.063548] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 901.063730] env[63021]: INFO nova.compute.manager [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Took 1.19 seconds to destroy the instance on the hypervisor. [ 901.063988] env[63021]: DEBUG oslo.service.loopingcall [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 901.064209] env[63021]: DEBUG nova.compute.manager [-] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 901.064851] env[63021]: DEBUG nova.network.neutron [-] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 901.116410] env[63021]: DEBUG oslo_vmware.api [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]524d1ad1-826a-981b-aae7-db97e2ebf9f3, 'name': SearchDatastore_Task, 'duration_secs': 0.015729} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.117231] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7dbaa30f-9c1f-4752-b06e-78202cc4c804 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.123056] env[63021]: DEBUG oslo_vmware.api [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 901.123056] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52edc4a8-bef7-833d-1403-fa7e9215df33" [ 901.123056] env[63021]: _type = "Task" [ 901.123056] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.134285] env[63021]: DEBUG oslo_vmware.api [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52edc4a8-bef7-833d-1403-fa7e9215df33, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.195485] env[63021]: DEBUG nova.network.neutron [req-44951688-29d9-4e2c-b3f2-c3bf334ebedc req-8589408c-ca9a-4d7f-846a-d8b92ea8d8cb service nova] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 901.282167] env[63021]: DEBUG nova.scheduler.client.report [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 901.393378] env[63021]: DEBUG nova.network.neutron [req-44951688-29d9-4e2c-b3f2-c3bf334ebedc req-8589408c-ca9a-4d7f-846a-d8b92ea8d8cb service nova] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.404560] env[63021]: DEBUG oslo_vmware.api [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293892, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.101221} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.405097] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 901.406065] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ef6d119-a8ac-455a-9366-714c89a84231 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.430731] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] 4d712cc1-6202-407e-8767-c620a1ce0bba/4d712cc1-6202-407e-8767-c620a1ce0bba.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 901.431378] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-186589d3-166d-4038-9d7d-4228263632d2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.453357] env[63021]: DEBUG oslo_vmware.api [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Waiting for the task: (returnval){ [ 901.453357] env[63021]: value = "task-1293895" [ 901.453357] env[63021]: _type = "Task" [ 901.453357] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.463383] env[63021]: DEBUG oslo_vmware.api [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293895, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.553807] env[63021]: DEBUG oslo_concurrency.lockutils [None req-072e7799-003f-447d-a797-f869bab1b1e2 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Lock "80b857c4-d9dd-4483-970f-a0f5f4bcf173" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.024s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.637854] env[63021]: DEBUG oslo_vmware.api [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52edc4a8-bef7-833d-1403-fa7e9215df33, 'name': SearchDatastore_Task, 'duration_secs': 0.017497} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.637854] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.637854] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 1b4871cc-3dac-4683-8fce-93c4d9e4407d/e624225d-af0f-415c-90cc-d4b228c479f6-rescue.vmdk. {{(pid=63021) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 901.637854] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8ae9272a-cca3-45d7-af77-cd8664ac9a3f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.645896] env[63021]: DEBUG oslo_vmware.api [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 901.645896] env[63021]: value = "task-1293896" [ 901.645896] env[63021]: _type = "Task" [ 901.645896] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.656072] env[63021]: DEBUG oslo_vmware.api [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293896, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.792225] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.931s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.794550] env[63021]: DEBUG oslo_concurrency.lockutils [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.085s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.796478] env[63021]: INFO nova.compute.claims [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 901.821042] env[63021]: INFO nova.scheduler.client.report [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Deleted allocations for instance 16b21798-17e8-4aeb-affa-57eae31d5dd3 [ 901.891656] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "refresh_cache-71909213-258d-45f5-9a3e-e473e8ce8aa2" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.891940] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquired lock "refresh_cache-71909213-258d-45f5-9a3e-e473e8ce8aa2" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.892137] env[63021]: DEBUG nova.network.neutron [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 901.906113] env[63021]: DEBUG oslo_concurrency.lockutils [req-44951688-29d9-4e2c-b3f2-c3bf334ebedc req-8589408c-ca9a-4d7f-846a-d8b92ea8d8cb service nova] Releasing lock "refresh_cache-ed26dff1-fed0-4baf-ad41-d14850254aec" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.907079] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Acquired lock "refresh_cache-ed26dff1-fed0-4baf-ad41-d14850254aec" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.907079] env[63021]: DEBUG nova.network.neutron [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 901.969467] env[63021]: DEBUG oslo_vmware.api [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293895, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.114087] env[63021]: DEBUG nova.network.neutron [-] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.158283] env[63021]: DEBUG oslo_vmware.api [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293896, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.335715] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e54609d6-8720-410d-a9a4-0ed042c4cd9d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "16b21798-17e8-4aeb-affa-57eae31d5dd3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.254s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.470029] env[63021]: DEBUG oslo_vmware.api [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293895, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.619691] env[63021]: INFO nova.compute.manager [-] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Took 1.56 seconds to deallocate network for instance. [ 902.664651] env[63021]: DEBUG oslo_vmware.api [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293896, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.667363] env[63021]: DEBUG nova.network.neutron [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 902.684414] env[63021]: DEBUG nova.compute.manager [req-6d744478-3c8a-4d08-a56f-48a96f138f92 req-2510f230-2a3d-4d91-bf5d-673b02e964b2 service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Received event network-vif-deleted-37dcb3f1-2931-40c6-96ba-b31b8fa1a72b {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 902.982718] env[63021]: DEBUG oslo_vmware.api [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293895, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.055186] env[63021]: DEBUG nova.network.neutron [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Updating instance_info_cache with network_info: [{"id": "2e8a7fc6-825c-45a5-b41c-d28410061aa3", "address": "fa:16:3e:61:15:99", "network": {"id": "f18b6fd7-dc66-48f7-a183-e22056ae2daf", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-748006433-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9a23f10cf5a4b178c7efff86041ce21", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e8a7fc6-82", "ovs_interfaceid": "2e8a7fc6-825c-45a5-b41c-d28410061aa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.119123] env[63021]: DEBUG nova.network.neutron [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Updating instance_info_cache with network_info: [{"id": "1019be13-e855-4eab-8201-a2ae2b22476c", "address": "fa:16:3e:32:70:89", "network": {"id": "708d0dc3-c651-4001-9638-1d3538ff76c9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-983471056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "805af233efa147b3982fb794cd9e6f73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1019be13-e8", "ovs_interfaceid": "1019be13-e855-4eab-8201-a2ae2b22476c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.130635] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-133bea3d-1ef9-41f5-86bd-e595cc98a21b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.136685] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.144743] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea369fdd-16ad-4e67-b71e-2f558787c60f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.181131] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe5b0642-b7de-4d45-b923-9bf3d8837466 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.187314] env[63021]: DEBUG oslo_vmware.api [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293896, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.032433} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.191017] env[63021]: INFO nova.virt.vmwareapi.ds_util [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 1b4871cc-3dac-4683-8fce-93c4d9e4407d/e624225d-af0f-415c-90cc-d4b228c479f6-rescue.vmdk. [ 903.191017] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99d6d42a-6157-469c-ae7e-0b00590c61d5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.196132] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68ce5ad6-a1fe-4598-ac41-5909933896f5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.221378] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] 1b4871cc-3dac-4683-8fce-93c4d9e4407d/e624225d-af0f-415c-90cc-d4b228c479f6-rescue.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 903.223686] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-adae9705-4504-4ba0-affa-44885297b12f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.245905] env[63021]: DEBUG nova.compute.provider_tree [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 903.260163] env[63021]: DEBUG oslo_vmware.api [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 903.260163] env[63021]: value = "task-1293898" [ 903.260163] env[63021]: _type = "Task" [ 903.260163] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.283651] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "140184cd-6e4c-425c-8b17-361a1f565e93" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.283893] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "140184cd-6e4c-425c-8b17-361a1f565e93" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.472989] env[63021]: DEBUG oslo_vmware.api [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293895, 'name': ReconfigVM_Task, 'duration_secs': 1.913021} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.473326] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Reconfigured VM instance instance-0000004e to attach disk [datastore1] 4d712cc1-6202-407e-8767-c620a1ce0bba/4d712cc1-6202-407e-8767-c620a1ce0bba.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 903.473970] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0779d6b6-48ff-4636-91d2-1608b5ce3367 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.480834] env[63021]: DEBUG oslo_vmware.api [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Waiting for the task: (returnval){ [ 903.480834] env[63021]: value = "task-1293899" [ 903.480834] env[63021]: _type = "Task" [ 903.480834] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.490801] env[63021]: DEBUG oslo_vmware.api [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293899, 'name': Rename_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.557856] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Releasing lock "refresh_cache-ed26dff1-fed0-4baf-ad41-d14850254aec" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.558211] env[63021]: DEBUG nova.compute.manager [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Instance network_info: |[{"id": "2e8a7fc6-825c-45a5-b41c-d28410061aa3", "address": "fa:16:3e:61:15:99", "network": {"id": "f18b6fd7-dc66-48f7-a183-e22056ae2daf", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-748006433-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9a23f10cf5a4b178c7efff86041ce21", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e8a7fc6-82", "ovs_interfaceid": "2e8a7fc6-825c-45a5-b41c-d28410061aa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 903.558712] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:61:15:99', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ad4fcde7-8926-402a-a9b7-4878d2bc1cf6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2e8a7fc6-825c-45a5-b41c-d28410061aa3', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 903.566560] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Creating folder: Project (c9a23f10cf5a4b178c7efff86041ce21). Parent ref: group-v277447. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 903.566857] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-612b95d8-7df2-4ee4-af29-cc2435d7a38b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.577100] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Created folder: Project (c9a23f10cf5a4b178c7efff86041ce21) in parent group-v277447. [ 903.577324] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Creating folder: Instances. Parent ref: group-v277541. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 903.577629] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d1a60773-c66c-4cbf-a150-c36f70da3c19 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.586033] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Created folder: Instances in parent group-v277541. [ 903.586225] env[63021]: DEBUG oslo.service.loopingcall [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 903.586431] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 903.586663] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e33e6072-85fe-4074-aa23-af94f0bfd375 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.605639] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 903.605639] env[63021]: value = "task-1293902" [ 903.605639] env[63021]: _type = "Task" [ 903.605639] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.613088] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293902, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.620792] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Releasing lock "refresh_cache-71909213-258d-45f5-9a3e-e473e8ce8aa2" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.749805] env[63021]: DEBUG nova.scheduler.client.report [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 903.770792] env[63021]: DEBUG oslo_vmware.api [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293898, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.786369] env[63021]: DEBUG nova.compute.manager [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 903.992742] env[63021]: DEBUG oslo_vmware.api [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293899, 'name': Rename_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.119761] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293902, 'name': CreateVM_Task} progress is 25%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.150044] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a05ed3d-8c83-4495-8924-bbe82254d026 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.170839] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e54914d-01cd-4912-820b-3983bd134d5d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.178609] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Updating instance '71909213-258d-45f5-9a3e-e473e8ce8aa2' progress to 83 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 904.257020] env[63021]: DEBUG oslo_concurrency.lockutils [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.460s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.257020] env[63021]: DEBUG nova.compute.manager [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 904.258477] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 8.507s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.258861] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.259186] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63021) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 904.259570] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.322s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.259880] env[63021]: DEBUG nova.objects.instance [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Lazy-loading 'resources' on Instance uuid f0154308-e6a8-4321-a082-99291344664e {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 904.261814] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7ddae1f-0921-4a56-afb3-d52a5475a94d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.277145] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aa26e6e-bdf5-4c2c-9563-ef37eac76a93 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.282547] env[63021]: DEBUG oslo_vmware.api [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293898, 'name': ReconfigVM_Task, 'duration_secs': 0.948876} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.282998] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Reconfigured VM instance instance-0000004c to attach disk [datastore2] 1b4871cc-3dac-4683-8fce-93c4d9e4407d/e624225d-af0f-415c-90cc-d4b228c479f6-rescue.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 904.284951] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db9b5f25-fdd2-4d5a-8049-511ad50da402 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.308020] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b7698b0-caa7-4d0f-91e4-1b3a614aa62f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.332278] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6b5acaa9-eb1d-4748-8560-a674a7ec140a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.343936] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.347414] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aaab2ec-a81b-4302-b22e-5417c758630d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.351453] env[63021]: DEBUG oslo_vmware.api [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 904.351453] env[63021]: value = "task-1293903" [ 904.351453] env[63021]: _type = "Task" [ 904.351453] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.380246] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180719MB free_disk=138GB free_vcpus=48 pci_devices=None {{(pid=63021) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 904.380447] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.385501] env[63021]: DEBUG oslo_vmware.api [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293903, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.491483] env[63021]: DEBUG oslo_vmware.api [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293899, 'name': Rename_Task, 'duration_secs': 0.568801} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.491827] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 904.492166] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-be4ad1af-e66b-4b58-ae06-a39cab07642b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.498539] env[63021]: DEBUG oslo_vmware.api [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Waiting for the task: (returnval){ [ 904.498539] env[63021]: value = "task-1293904" [ 904.498539] env[63021]: _type = "Task" [ 904.498539] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.506347] env[63021]: DEBUG oslo_vmware.api [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293904, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.617995] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293902, 'name': CreateVM_Task, 'duration_secs': 0.731225} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.618214] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 904.618930] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.619115] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.619442] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 904.619729] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca937fc2-8784-4c1f-98b7-4f6cf25ecfbb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.624544] env[63021]: DEBUG oslo_vmware.api [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Waiting for the task: (returnval){ [ 904.624544] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52ea2bb6-78c2-f432-128d-e5c7f7d5007d" [ 904.624544] env[63021]: _type = "Task" [ 904.624544] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.632427] env[63021]: DEBUG oslo_vmware.api [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52ea2bb6-78c2-f432-128d-e5c7f7d5007d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.685846] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 904.686226] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f6967c4d-9d0c-4cab-b7e7-6be0563645cd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.694170] env[63021]: DEBUG oslo_vmware.api [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 904.694170] env[63021]: value = "task-1293905" [ 904.694170] env[63021]: _type = "Task" [ 904.694170] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.705209] env[63021]: DEBUG oslo_vmware.api [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1293905, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.714549] env[63021]: DEBUG nova.compute.manager [req-ffdb55ad-bb0c-4c0e-9bf7-96efee20c6bd req-d14d195b-1e1f-4c21-a317-4f06c4f05d26 service nova] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Received event network-changed-3dc5a6b8-23fe-42fa-88c1-6854a326150d {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 904.714745] env[63021]: DEBUG nova.compute.manager [req-ffdb55ad-bb0c-4c0e-9bf7-96efee20c6bd req-d14d195b-1e1f-4c21-a317-4f06c4f05d26 service nova] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Refreshing instance network info cache due to event network-changed-3dc5a6b8-23fe-42fa-88c1-6854a326150d. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 904.714974] env[63021]: DEBUG oslo_concurrency.lockutils [req-ffdb55ad-bb0c-4c0e-9bf7-96efee20c6bd req-d14d195b-1e1f-4c21-a317-4f06c4f05d26 service nova] Acquiring lock "refresh_cache-80b857c4-d9dd-4483-970f-a0f5f4bcf173" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.715184] env[63021]: DEBUG oslo_concurrency.lockutils [req-ffdb55ad-bb0c-4c0e-9bf7-96efee20c6bd req-d14d195b-1e1f-4c21-a317-4f06c4f05d26 service nova] Acquired lock "refresh_cache-80b857c4-d9dd-4483-970f-a0f5f4bcf173" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.715359] env[63021]: DEBUG nova.network.neutron [req-ffdb55ad-bb0c-4c0e-9bf7-96efee20c6bd req-d14d195b-1e1f-4c21-a317-4f06c4f05d26 service nova] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Refreshing network info cache for port 3dc5a6b8-23fe-42fa-88c1-6854a326150d {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 904.762525] env[63021]: DEBUG nova.compute.utils [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 904.763818] env[63021]: DEBUG nova.compute.manager [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 904.763977] env[63021]: DEBUG nova.network.neutron [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 904.829887] env[63021]: DEBUG nova.policy [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1e133aeda3249ec98dbdc53f869d719', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7331caed42034dfc81c79cfabbe0829e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 904.861972] env[63021]: DEBUG oslo_vmware.api [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293903, 'name': ReconfigVM_Task, 'duration_secs': 0.318254} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.864601] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 904.865646] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0a254b5e-24cd-4b6a-8d24-0290e8ca7b10 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.874511] env[63021]: DEBUG oslo_vmware.api [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 904.874511] env[63021]: value = "task-1293906" [ 904.874511] env[63021]: _type = "Task" [ 904.874511] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.884465] env[63021]: DEBUG oslo_vmware.api [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293906, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.014709] env[63021]: DEBUG oslo_vmware.api [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293904, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.101664] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f8d8b85-c817-4dc3-9a91-63dab335887d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.110910] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bf02816-2b8d-411a-8c8b-03d76396bee2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.151830] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e24d312-1626-4e14-a999-9562a1a423b8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.155696] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-22810790-0d2c-4212-8068-81258f676e07 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Volume attach. Driver type: vmdk {{(pid=63021) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 905.155999] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-22810790-0d2c-4212-8068-81258f676e07 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277540', 'volume_id': '1d748e34-0bec-4d79-9a3a-f42d78520ff7', 'name': 'volume-1d748e34-0bec-4d79-9a3a-f42d78520ff7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'df856146-61ba-4de7-a5fd-d75263927163', 'attached_at': '', 'detached_at': '', 'volume_id': '1d748e34-0bec-4d79-9a3a-f42d78520ff7', 'serial': '1d748e34-0bec-4d79-9a3a-f42d78520ff7'} {{(pid=63021) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 905.156859] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-994ffdce-544a-4296-8b2c-13e7a0e8caf3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.160447] env[63021]: DEBUG nova.network.neutron [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Successfully created port: eb6c1950-a21a-4873-9c10-53065bf84e01 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 905.167915] env[63021]: DEBUG oslo_vmware.api [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52ea2bb6-78c2-f432-128d-e5c7f7d5007d, 'name': SearchDatastore_Task, 'duration_secs': 0.018414} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.180735] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad091d3c-e0d8-4ecb-a839-ab0781ba5324 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.184915] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.185219] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 905.185469] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.185618] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.185800] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 905.186251] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-faf722df-0c01-4e8c-bb9a-ff37b548d357 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.188466] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-644babc1-1043-4938-bbd4-8196d878e72d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.201495] env[63021]: DEBUG nova.compute.provider_tree [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 905.220853] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 905.221073] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 905.231681] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-22810790-0d2c-4212-8068-81258f676e07 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] volume-1d748e34-0bec-4d79-9a3a-f42d78520ff7/volume-1d748e34-0bec-4d79-9a3a-f42d78520ff7.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 905.232907] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0a678c5-429f-4a6f-91cf-97b35bdedc94 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.235611] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-963bb982-3b98-4e59-9b76-02525588d5a7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.254722] env[63021]: DEBUG oslo_vmware.api [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1293905, 'name': PowerOnVM_Task, 'duration_secs': 0.452154} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.257057] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 905.257274] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-fa9d0aca-1123-4e4c-98c7-aa4bdc8c5a88 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Updating instance '71909213-258d-45f5-9a3e-e473e8ce8aa2' progress to 100 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 905.261567] env[63021]: DEBUG oslo_vmware.api [None req-22810790-0d2c-4212-8068-81258f676e07 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 905.261567] env[63021]: value = "task-1293907" [ 905.261567] env[63021]: _type = "Task" [ 905.261567] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.261879] env[63021]: DEBUG oslo_vmware.api [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Waiting for the task: (returnval){ [ 905.261879] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52b347fe-8711-4648-1c69-a6718daa65e6" [ 905.261879] env[63021]: _type = "Task" [ 905.261879] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.269562] env[63021]: DEBUG nova.compute.manager [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 905.280900] env[63021]: DEBUG oslo_vmware.api [None req-22810790-0d2c-4212-8068-81258f676e07 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1293907, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.285385] env[63021]: DEBUG oslo_vmware.api [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52b347fe-8711-4648-1c69-a6718daa65e6, 'name': SearchDatastore_Task, 'duration_secs': 0.019327} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.285898] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-165fc27e-d53c-44a3-a28e-8a6facd8a012 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.292330] env[63021]: DEBUG oslo_vmware.api [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Waiting for the task: (returnval){ [ 905.292330] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]520fb18b-61d1-7eeb-0ac9-bc64125b0fc0" [ 905.292330] env[63021]: _type = "Task" [ 905.292330] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.302122] env[63021]: DEBUG oslo_vmware.api [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]520fb18b-61d1-7eeb-0ac9-bc64125b0fc0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.385776] env[63021]: DEBUG oslo_vmware.api [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293906, 'name': PowerOnVM_Task, 'duration_secs': 0.469007} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.386608] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 905.389291] env[63021]: DEBUG nova.compute.manager [None req-9a960a32-a9ac-47c1-a089-86119a1b7e62 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 905.390656] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aecdc07-3ac8-436a-8b6c-781c87dcb657 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.511199] env[63021]: DEBUG oslo_vmware.api [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293904, 'name': PowerOnVM_Task, 'duration_secs': 0.651636} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.515099] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 905.515099] env[63021]: INFO nova.compute.manager [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Took 7.95 seconds to spawn the instance on the hypervisor. [ 905.515099] env[63021]: DEBUG nova.compute.manager [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 905.515099] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8706e75e-d577-412e-9a88-d8fa4c4552ad {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.675521] env[63021]: DEBUG nova.network.neutron [req-ffdb55ad-bb0c-4c0e-9bf7-96efee20c6bd req-d14d195b-1e1f-4c21-a317-4f06c4f05d26 service nova] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Updated VIF entry in instance network info cache for port 3dc5a6b8-23fe-42fa-88c1-6854a326150d. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 905.675521] env[63021]: DEBUG nova.network.neutron [req-ffdb55ad-bb0c-4c0e-9bf7-96efee20c6bd req-d14d195b-1e1f-4c21-a317-4f06c4f05d26 service nova] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Updating instance_info_cache with network_info: [{"id": "3dc5a6b8-23fe-42fa-88c1-6854a326150d", "address": "fa:16:3e:13:6a:e4", "network": {"id": "1a4029b0-70bb-49c1-9689-7517cf4b29d3", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-875049177-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.232", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52e7242ba61e4ce28c8f3b792980bdac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f762954-6ca5-4da5-bf0a-5d31c51ec570", "external-id": "nsx-vlan-transportzone-930", "segmentation_id": 930, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3dc5a6b8-23", "ovs_interfaceid": "3dc5a6b8-23fe-42fa-88c1-6854a326150d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.706273] env[63021]: DEBUG nova.scheduler.client.report [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 905.773131] env[63021]: DEBUG oslo_vmware.api [None req-22810790-0d2c-4212-8068-81258f676e07 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1293907, 'name': ReconfigVM_Task, 'duration_secs': 0.424198} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.777019] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-22810790-0d2c-4212-8068-81258f676e07 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Reconfigured VM instance instance-00000043 to attach disk [datastore2] volume-1d748e34-0bec-4d79-9a3a-f42d78520ff7/volume-1d748e34-0bec-4d79-9a3a-f42d78520ff7.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 905.782206] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dd569d18-7fce-4ea7-a5c3-09146aada9c2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.808028] env[63021]: DEBUG oslo_vmware.api [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]520fb18b-61d1-7eeb-0ac9-bc64125b0fc0, 'name': SearchDatastore_Task, 'duration_secs': 0.010682} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.808028] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.808028] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] ed26dff1-fed0-4baf-ad41-d14850254aec/ed26dff1-fed0-4baf-ad41-d14850254aec.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 905.808028] env[63021]: DEBUG oslo_vmware.api [None req-22810790-0d2c-4212-8068-81258f676e07 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 905.808028] env[63021]: value = "task-1293908" [ 905.808028] env[63021]: _type = "Task" [ 905.808028] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.808028] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c9ce8953-4da2-4ae0-bf26-b8ca3c0c5dc5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.818537] env[63021]: DEBUG oslo_vmware.api [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Waiting for the task: (returnval){ [ 905.818537] env[63021]: value = "task-1293909" [ 905.818537] env[63021]: _type = "Task" [ 905.818537] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.822542] env[63021]: DEBUG oslo_vmware.api [None req-22810790-0d2c-4212-8068-81258f676e07 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1293908, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.831536] env[63021]: DEBUG oslo_vmware.api [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Task: {'id': task-1293909, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.035791] env[63021]: INFO nova.compute.manager [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Took 38.62 seconds to build instance. [ 906.178285] env[63021]: DEBUG oslo_concurrency.lockutils [req-ffdb55ad-bb0c-4c0e-9bf7-96efee20c6bd req-d14d195b-1e1f-4c21-a317-4f06c4f05d26 service nova] Releasing lock "refresh_cache-80b857c4-d9dd-4483-970f-a0f5f4bcf173" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.213889] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.954s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.216960] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bfd9b1b2-3da1-4969-9916-ce226c3b8037 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.719s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.217533] env[63021]: DEBUG nova.objects.instance [None req-bfd9b1b2-3da1-4969-9916-ce226c3b8037 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Lazy-loading 'resources' on Instance uuid 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 906.244944] env[63021]: INFO nova.scheduler.client.report [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Deleted allocations for instance f0154308-e6a8-4321-a082-99291344664e [ 906.295397] env[63021]: DEBUG nova.compute.manager [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 906.336922] env[63021]: DEBUG oslo_vmware.api [None req-22810790-0d2c-4212-8068-81258f676e07 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1293908, 'name': ReconfigVM_Task, 'duration_secs': 0.141317} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.341975] env[63021]: DEBUG nova.virt.hardware [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='b662fa2ed1ce2ebd193dbdf728cfb4e3',container_format='bare',created_at=2024-10-15T17:49:41Z,direct_url=,disk_format='vmdk',id=02de5fc5-60f6-49e2-ae03-e14c840dae1b,min_disk=1,min_ram=0,name='tempest-test-snap-789413026',owner='7331caed42034dfc81c79cfabbe0829e',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-15T17:49:58Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 906.342155] env[63021]: DEBUG nova.virt.hardware [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 906.342355] env[63021]: DEBUG nova.virt.hardware [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 906.342635] env[63021]: DEBUG nova.virt.hardware [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 906.342862] env[63021]: DEBUG nova.virt.hardware [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 906.343680] env[63021]: DEBUG nova.virt.hardware [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 906.343788] env[63021]: DEBUG nova.virt.hardware [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 906.343942] env[63021]: DEBUG nova.virt.hardware [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 906.344234] env[63021]: DEBUG nova.virt.hardware [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 906.344524] env[63021]: DEBUG nova.virt.hardware [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 906.344754] env[63021]: DEBUG nova.virt.hardware [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 906.346079] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-22810790-0d2c-4212-8068-81258f676e07 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277540', 'volume_id': '1d748e34-0bec-4d79-9a3a-f42d78520ff7', 'name': 'volume-1d748e34-0bec-4d79-9a3a-f42d78520ff7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'df856146-61ba-4de7-a5fd-d75263927163', 'attached_at': '', 'detached_at': '', 'volume_id': '1d748e34-0bec-4d79-9a3a-f42d78520ff7', 'serial': '1d748e34-0bec-4d79-9a3a-f42d78520ff7'} {{(pid=63021) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 906.356205] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e3657c1-9fd5-4fb0-a808-b084f5922c4c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.371921] env[63021]: DEBUG oslo_vmware.api [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Task: {'id': task-1293909, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.381436] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec569ce9-be40-446d-a4bc-35cf9d9a8df5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.539876] env[63021]: DEBUG oslo_concurrency.lockutils [None req-27fe901d-3b86-4ae4-8518-a223d9c72379 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Lock "4d712cc1-6202-407e-8767-c620a1ce0bba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 58.548s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.577549] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Acquiring lock "45c9e2be-eb19-4b83-b280-c9eeaddfccba" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.577881] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lock "45c9e2be-eb19-4b83-b280-c9eeaddfccba" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.578086] env[63021]: INFO nova.compute.manager [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Shelving [ 906.652337] env[63021]: INFO nova.compute.manager [None req-64b4d8c9-e71b-42cc-8354-aabb9a08ff12 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Unrescuing [ 906.654180] env[63021]: DEBUG oslo_concurrency.lockutils [None req-64b4d8c9-e71b-42cc-8354-aabb9a08ff12 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquiring lock "refresh_cache-1b4871cc-3dac-4683-8fce-93c4d9e4407d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.654180] env[63021]: DEBUG oslo_concurrency.lockutils [None req-64b4d8c9-e71b-42cc-8354-aabb9a08ff12 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquired lock "refresh_cache-1b4871cc-3dac-4683-8fce-93c4d9e4407d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.654180] env[63021]: DEBUG nova.network.neutron [None req-64b4d8c9-e71b-42cc-8354-aabb9a08ff12 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 906.753839] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0a110bd8-dc6d-4b96-b5d6-6298e5c227a2 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Lock "f0154308-e6a8-4321-a082-99291344664e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.747s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.833523] env[63021]: DEBUG oslo_vmware.api [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Task: {'id': task-1293909, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.879782] env[63021]: INFO nova.compute.manager [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Rebuilding instance [ 906.933467] env[63021]: DEBUG nova.compute.manager [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 906.934365] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3c99e83-90d5-4a50-9a05-978a351f964c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.957515] env[63021]: DEBUG nova.network.neutron [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Successfully updated port: eb6c1950-a21a-4873-9c10-53065bf84e01 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 907.024256] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d99b2f1-ba04-44c6-9ec2-b34eb93fc352 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.032082] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-645634ff-3f2a-4642-9936-cce27da156b0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.062952] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-850282e2-c775-4886-9aed-e5f2a28f405f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.070637] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-648606ae-1deb-4347-9896-8983e7ee934b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.078669] env[63021]: DEBUG nova.compute.manager [req-27bb8cba-ab1a-456e-904a-61cd28ae93cc req-19127d75-2689-429a-ade0-ce59553a8e8a service nova] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Received event network-vif-plugged-eb6c1950-a21a-4873-9c10-53065bf84e01 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 907.078669] env[63021]: DEBUG oslo_concurrency.lockutils [req-27bb8cba-ab1a-456e-904a-61cd28ae93cc req-19127d75-2689-429a-ade0-ce59553a8e8a service nova] Acquiring lock "d63fd465-975f-42ce-b1be-ac8a6929bc4d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.078669] env[63021]: DEBUG oslo_concurrency.lockutils [req-27bb8cba-ab1a-456e-904a-61cd28ae93cc req-19127d75-2689-429a-ade0-ce59553a8e8a service nova] Lock "d63fd465-975f-42ce-b1be-ac8a6929bc4d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.078669] env[63021]: DEBUG oslo_concurrency.lockutils [req-27bb8cba-ab1a-456e-904a-61cd28ae93cc req-19127d75-2689-429a-ade0-ce59553a8e8a service nova] Lock "d63fd465-975f-42ce-b1be-ac8a6929bc4d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.079083] env[63021]: DEBUG nova.compute.manager [req-27bb8cba-ab1a-456e-904a-61cd28ae93cc req-19127d75-2689-429a-ade0-ce59553a8e8a service nova] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] No waiting events found dispatching network-vif-plugged-eb6c1950-a21a-4873-9c10-53065bf84e01 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 907.079083] env[63021]: WARNING nova.compute.manager [req-27bb8cba-ab1a-456e-904a-61cd28ae93cc req-19127d75-2689-429a-ade0-ce59553a8e8a service nova] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Received unexpected event network-vif-plugged-eb6c1950-a21a-4873-9c10-53065bf84e01 for instance with vm_state building and task_state spawning. [ 907.090769] env[63021]: DEBUG nova.compute.provider_tree [None req-bfd9b1b2-3da1-4969-9916-ce226c3b8037 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 907.093853] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 907.094302] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-91f97da0-07a8-4845-badc-1e90f31359d1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.100671] env[63021]: DEBUG oslo_vmware.api [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Waiting for the task: (returnval){ [ 907.100671] env[63021]: value = "task-1293910" [ 907.100671] env[63021]: _type = "Task" [ 907.100671] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.110104] env[63021]: DEBUG oslo_vmware.api [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1293910, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.182178] env[63021]: DEBUG oslo_concurrency.lockutils [None req-65a46cd5-0e5c-4e9c-bb88-b7b7e09d89a7 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "71909213-258d-45f5-9a3e-e473e8ce8aa2" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.182492] env[63021]: DEBUG oslo_concurrency.lockutils [None req-65a46cd5-0e5c-4e9c-bb88-b7b7e09d89a7 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "71909213-258d-45f5-9a3e-e473e8ce8aa2" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.182767] env[63021]: DEBUG nova.compute.manager [None req-65a46cd5-0e5c-4e9c-bb88-b7b7e09d89a7 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Going to confirm migration 1 {{(pid=63021) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 907.334449] env[63021]: DEBUG oslo_vmware.api [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Task: {'id': task-1293909, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.390441] env[63021]: DEBUG nova.network.neutron [None req-64b4d8c9-e71b-42cc-8354-aabb9a08ff12 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Updating instance_info_cache with network_info: [{"id": "d34cfd0d-464e-400b-8366-83321cc6112f", "address": "fa:16:3e:64:67:d0", "network": {"id": "37008b1f-68ff-404c-9afc-14b108af85c3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1473182647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "2d364c9eae4b4f8a8b60d4a3ac2ed648", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "02bbcead-d833-4543-bec6-fb82dfe659ff", "external-id": "nsx-vlan-transportzone-478", "segmentation_id": 478, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd34cfd0d-46", "ovs_interfaceid": "d34cfd0d-464e-400b-8366-83321cc6112f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.420754] env[63021]: DEBUG nova.objects.instance [None req-22810790-0d2c-4212-8068-81258f676e07 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lazy-loading 'flavor' on Instance uuid df856146-61ba-4de7-a5fd-d75263927163 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 907.456801] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 907.457415] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7d170f2c-ba1f-42bb-880d-f21f33d17605 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.462019] env[63021]: DEBUG oslo_concurrency.lockutils [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "refresh_cache-d63fd465-975f-42ce-b1be-ac8a6929bc4d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.462211] env[63021]: DEBUG oslo_concurrency.lockutils [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquired lock "refresh_cache-d63fd465-975f-42ce-b1be-ac8a6929bc4d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.462380] env[63021]: DEBUG nova.network.neutron [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 907.467200] env[63021]: DEBUG oslo_vmware.api [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Waiting for the task: (returnval){ [ 907.467200] env[63021]: value = "task-1293911" [ 907.467200] env[63021]: _type = "Task" [ 907.467200] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.475908] env[63021]: DEBUG oslo_vmware.api [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293911, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.597792] env[63021]: DEBUG nova.scheduler.client.report [None req-bfd9b1b2-3da1-4969-9916-ce226c3b8037 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 907.616361] env[63021]: DEBUG oslo_vmware.api [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1293910, 'name': PowerOffVM_Task, 'duration_secs': 0.214816} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.616793] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 907.618543] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d71f181d-844f-4cbf-b784-445cb10c6681 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.643963] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2175ab0d-2741-401d-96db-fea97be3e7a1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.798512] env[63021]: DEBUG oslo_concurrency.lockutils [None req-65a46cd5-0e5c-4e9c-bb88-b7b7e09d89a7 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "refresh_cache-71909213-258d-45f5-9a3e-e473e8ce8aa2" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.798693] env[63021]: DEBUG oslo_concurrency.lockutils [None req-65a46cd5-0e5c-4e9c-bb88-b7b7e09d89a7 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquired lock "refresh_cache-71909213-258d-45f5-9a3e-e473e8ce8aa2" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.798871] env[63021]: DEBUG nova.network.neutron [None req-65a46cd5-0e5c-4e9c-bb88-b7b7e09d89a7 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 907.799076] env[63021]: DEBUG nova.objects.instance [None req-65a46cd5-0e5c-4e9c-bb88-b7b7e09d89a7 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lazy-loading 'info_cache' on Instance uuid 71909213-258d-45f5-9a3e-e473e8ce8aa2 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 907.818219] env[63021]: DEBUG oslo_concurrency.lockutils [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "ae103118-bddf-46fe-90b9-98b60952ebba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.818219] env[63021]: DEBUG oslo_concurrency.lockutils [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "ae103118-bddf-46fe-90b9-98b60952ebba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.836952] env[63021]: DEBUG oslo_vmware.api [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Task: {'id': task-1293909, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.668901} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.837287] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] ed26dff1-fed0-4baf-ad41-d14850254aec/ed26dff1-fed0-4baf-ad41-d14850254aec.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 907.837718] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 907.838202] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-43676fe6-95ea-44b8-af6a-422015697271 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.846127] env[63021]: DEBUG oslo_vmware.api [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Waiting for the task: (returnval){ [ 907.846127] env[63021]: value = "task-1293912" [ 907.846127] env[63021]: _type = "Task" [ 907.846127] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.855148] env[63021]: DEBUG oslo_vmware.api [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Task: {'id': task-1293912, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.893390] env[63021]: DEBUG oslo_concurrency.lockutils [None req-64b4d8c9-e71b-42cc-8354-aabb9a08ff12 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Releasing lock "refresh_cache-1b4871cc-3dac-4683-8fce-93c4d9e4407d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.894104] env[63021]: DEBUG nova.objects.instance [None req-64b4d8c9-e71b-42cc-8354-aabb9a08ff12 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Lazy-loading 'flavor' on Instance uuid 1b4871cc-3dac-4683-8fce-93c4d9e4407d {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 907.926286] env[63021]: DEBUG oslo_concurrency.lockutils [None req-22810790-0d2c-4212-8068-81258f676e07 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "df856146-61ba-4de7-a5fd-d75263927163" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.435s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.979163] env[63021]: DEBUG oslo_vmware.api [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293911, 'name': PowerOffVM_Task, 'duration_secs': 0.313339} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.979524] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 907.979776] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 907.981059] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d519aa6-3af0-4c2d-b211-25119e58a812 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.989436] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 907.989697] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-535e3cb1-a249-4715-be6c-5e26eabd02d8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.007241] env[63021]: DEBUG nova.network.neutron [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 908.012742] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 908.013332] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 908.013572] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Deleting the datastore file [datastore1] 4d712cc1-6202-407e-8767-c620a1ce0bba {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 908.013852] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3e4a334f-218c-4042-a419-a6ef8cfb8368 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.020423] env[63021]: DEBUG oslo_vmware.api [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Waiting for the task: (returnval){ [ 908.020423] env[63021]: value = "task-1293914" [ 908.020423] env[63021]: _type = "Task" [ 908.020423] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.030392] env[63021]: DEBUG oslo_vmware.api [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293914, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.105089] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bfd9b1b2-3da1-4969-9916-ce226c3b8037 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.888s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.111086] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.971s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.111086] env[63021]: DEBUG nova.objects.instance [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lazy-loading 'resources' on Instance uuid 36ec45ec-761b-4d62-a74f-e4d9a840ada0 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 908.140893] env[63021]: INFO nova.scheduler.client.report [None req-bfd9b1b2-3da1-4969-9916-ce226c3b8037 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Deleted allocations for instance 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb [ 908.156889] env[63021]: DEBUG oslo_concurrency.lockutils [None req-92ae3262-30a3-4d02-a8d7-c51a4282461f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquiring lock "df856146-61ba-4de7-a5fd-d75263927163" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.157196] env[63021]: DEBUG oslo_concurrency.lockutils [None req-92ae3262-30a3-4d02-a8d7-c51a4282461f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "df856146-61ba-4de7-a5fd-d75263927163" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.157392] env[63021]: DEBUG nova.compute.manager [None req-92ae3262-30a3-4d02-a8d7-c51a4282461f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 908.159485] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Creating Snapshot of the VM instance {{(pid=63021) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 908.159485] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fb6568f-e307-4d13-b543-fb0019f4e4fd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.162532] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-9ba2383e-dda3-42ac-aa0b-4ee38850eac3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.169601] env[63021]: DEBUG nova.compute.manager [None req-92ae3262-30a3-4d02-a8d7-c51a4282461f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63021) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 908.170752] env[63021]: DEBUG nova.objects.instance [None req-92ae3262-30a3-4d02-a8d7-c51a4282461f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lazy-loading 'flavor' on Instance uuid df856146-61ba-4de7-a5fd-d75263927163 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 908.175348] env[63021]: DEBUG oslo_vmware.api [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Waiting for the task: (returnval){ [ 908.175348] env[63021]: value = "task-1293915" [ 908.175348] env[63021]: _type = "Task" [ 908.175348] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.184753] env[63021]: DEBUG oslo_concurrency.lockutils [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "418a394b-dbb2-47a2-96a9-a1caa4c4bb51" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.184753] env[63021]: DEBUG oslo_concurrency.lockutils [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "418a394b-dbb2-47a2-96a9-a1caa4c4bb51" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.189821] env[63021]: DEBUG oslo_vmware.api [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1293915, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.197640] env[63021]: DEBUG nova.network.neutron [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Updating instance_info_cache with network_info: [{"id": "eb6c1950-a21a-4873-9c10-53065bf84e01", "address": "fa:16:3e:b9:50:07", "network": {"id": "265e2e2e-6ec6-4d4c-8b4c-0a2e8baa6bdb", "bridge": "br-int", "label": "tempest-ImagesTestJSON-242691215-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7331caed42034dfc81c79cfabbe0829e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bfae3ef8-cae7-455d-8632-ba93e1671625", "external-id": "cl2-zone-841", "segmentation_id": 841, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb6c1950-a2", "ovs_interfaceid": "eb6c1950-a21a-4873-9c10-53065bf84e01", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.324328] env[63021]: DEBUG nova.compute.manager [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 908.360346] env[63021]: DEBUG oslo_vmware.api [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Task: {'id': task-1293912, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.132464} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.360566] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 908.361393] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7d9cc85-7094-4fc9-97a9-f695cbd50fe9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.387996] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] ed26dff1-fed0-4baf-ad41-d14850254aec/ed26dff1-fed0-4baf-ad41-d14850254aec.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 908.388695] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9a047015-3827-4f37-b705-17c335716606 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.411749] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c14111a-d64f-452f-9735-0d066f506190 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.416628] env[63021]: DEBUG oslo_vmware.api [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Waiting for the task: (returnval){ [ 908.416628] env[63021]: value = "task-1293916" [ 908.416628] env[63021]: _type = "Task" [ 908.416628] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.441897] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-64b4d8c9-e71b-42cc-8354-aabb9a08ff12 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 908.443086] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a2ad1549-c9d5-40f4-b4c0-2ec8e64da7cf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.448322] env[63021]: DEBUG oslo_vmware.api [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Task: {'id': task-1293916, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.453413] env[63021]: DEBUG oslo_vmware.api [None req-64b4d8c9-e71b-42cc-8354-aabb9a08ff12 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 908.453413] env[63021]: value = "task-1293917" [ 908.453413] env[63021]: _type = "Task" [ 908.453413] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.462144] env[63021]: DEBUG oslo_vmware.api [None req-64b4d8c9-e71b-42cc-8354-aabb9a08ff12 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293917, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.535926] env[63021]: DEBUG oslo_vmware.api [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293914, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.230462} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.536633] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 908.536633] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 908.536761] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 908.652975] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bfd9b1b2-3da1-4969-9916-ce226c3b8037 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Lock "51e4a14c-4792-43cd-9e99-4e06a8f6d6fb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.787s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.682098] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-92ae3262-30a3-4d02-a8d7-c51a4282461f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 908.684932] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8f81e278-0ef4-4261-8134-e2a3467d65c9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.692393] env[63021]: DEBUG nova.compute.manager [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 908.695101] env[63021]: DEBUG oslo_vmware.api [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1293915, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.697199] env[63021]: DEBUG oslo_vmware.api [None req-92ae3262-30a3-4d02-a8d7-c51a4282461f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 908.697199] env[63021]: value = "task-1293918" [ 908.697199] env[63021]: _type = "Task" [ 908.697199] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.701118] env[63021]: DEBUG oslo_concurrency.lockutils [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Releasing lock "refresh_cache-d63fd465-975f-42ce-b1be-ac8a6929bc4d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.701118] env[63021]: DEBUG nova.compute.manager [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Instance network_info: |[{"id": "eb6c1950-a21a-4873-9c10-53065bf84e01", "address": "fa:16:3e:b9:50:07", "network": {"id": "265e2e2e-6ec6-4d4c-8b4c-0a2e8baa6bdb", "bridge": "br-int", "label": "tempest-ImagesTestJSON-242691215-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7331caed42034dfc81c79cfabbe0829e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bfae3ef8-cae7-455d-8632-ba93e1671625", "external-id": "cl2-zone-841", "segmentation_id": 841, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb6c1950-a2", "ovs_interfaceid": "eb6c1950-a21a-4873-9c10-53065bf84e01", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 908.705134] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b9:50:07', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bfae3ef8-cae7-455d-8632-ba93e1671625', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eb6c1950-a21a-4873-9c10-53065bf84e01', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 908.714095] env[63021]: DEBUG oslo.service.loopingcall [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 908.717840] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 908.717914] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-201de35c-d01e-495c-8c35-23a159862e7c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.742603] env[63021]: DEBUG oslo_vmware.api [None req-92ae3262-30a3-4d02-a8d7-c51a4282461f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1293918, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.747493] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 908.747493] env[63021]: value = "task-1293919" [ 908.747493] env[63021]: _type = "Task" [ 908.747493] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.755694] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293919, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.845720] env[63021]: DEBUG oslo_concurrency.lockutils [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.902761] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5e88368-0d69-4bb4-83b8-ef0619875fda {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.910231] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a9aecc0-8e6e-4038-a734-58d06776fb64 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.915784] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cb8a5d1d-bec9-4830-b36d-432c88f95588 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquiring lock "44dadf8e-a727-448f-887d-6408f7ebfbc6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.915905] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cb8a5d1d-bec9-4830-b36d-432c88f95588 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Lock "44dadf8e-a727-448f-887d-6408f7ebfbc6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.916182] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cb8a5d1d-bec9-4830-b36d-432c88f95588 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquiring lock "44dadf8e-a727-448f-887d-6408f7ebfbc6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.916182] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cb8a5d1d-bec9-4830-b36d-432c88f95588 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Lock "44dadf8e-a727-448f-887d-6408f7ebfbc6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.916775] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cb8a5d1d-bec9-4830-b36d-432c88f95588 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Lock "44dadf8e-a727-448f-887d-6408f7ebfbc6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.919104] env[63021]: INFO nova.compute.manager [None req-cb8a5d1d-bec9-4830-b36d-432c88f95588 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Terminating instance [ 909.640546] env[63021]: DEBUG nova.compute.manager [None req-cb8a5d1d-bec9-4830-b36d-432c88f95588 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 909.640790] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-cb8a5d1d-bec9-4830-b36d-432c88f95588 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 909.643728] env[63021]: DEBUG oslo_concurrency.lockutils [None req-95b6e88d-8134-48e8-a7c6-826dc81538bf tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquiring lock "50a1269e-4d83-4cf7-ba14-3b1afb3c43ed" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.643966] env[63021]: DEBUG oslo_concurrency.lockutils [None req-95b6e88d-8134-48e8-a7c6-826dc81538bf tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Lock "50a1269e-4d83-4cf7-ba14-3b1afb3c43ed" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.644202] env[63021]: DEBUG oslo_concurrency.lockutils [None req-95b6e88d-8134-48e8-a7c6-826dc81538bf tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquiring lock "50a1269e-4d83-4cf7-ba14-3b1afb3c43ed-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.644433] env[63021]: DEBUG oslo_concurrency.lockutils [None req-95b6e88d-8134-48e8-a7c6-826dc81538bf tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Lock "50a1269e-4d83-4cf7-ba14-3b1afb3c43ed-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.644631] env[63021]: DEBUG oslo_concurrency.lockutils [None req-95b6e88d-8134-48e8-a7c6-826dc81538bf tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Lock "50a1269e-4d83-4cf7-ba14-3b1afb3c43ed-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.654873] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9359d1e-cec3-4659-981b-76e809934ada {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.659472] env[63021]: INFO nova.compute.manager [None req-95b6e88d-8134-48e8-a7c6-826dc81538bf tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Terminating instance [ 909.666500] env[63021]: DEBUG nova.compute.manager [req-15e7829d-b603-40c4-b0fa-ebc9db5026b0 req-c9acd85a-5b4a-46d2-86ce-bbd4f1e435f0 service nova] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Received event network-changed-eb6c1950-a21a-4873-9c10-53065bf84e01 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 909.666500] env[63021]: DEBUG nova.compute.manager [req-15e7829d-b603-40c4-b0fa-ebc9db5026b0 req-c9acd85a-5b4a-46d2-86ce-bbd4f1e435f0 service nova] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Refreshing instance network info cache due to event network-changed-eb6c1950-a21a-4873-9c10-53065bf84e01. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 909.666500] env[63021]: DEBUG oslo_concurrency.lockutils [req-15e7829d-b603-40c4-b0fa-ebc9db5026b0 req-c9acd85a-5b4a-46d2-86ce-bbd4f1e435f0 service nova] Acquiring lock "refresh_cache-d63fd465-975f-42ce-b1be-ac8a6929bc4d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.666500] env[63021]: DEBUG oslo_concurrency.lockutils [req-15e7829d-b603-40c4-b0fa-ebc9db5026b0 req-c9acd85a-5b4a-46d2-86ce-bbd4f1e435f0 service nova] Acquired lock "refresh_cache-d63fd465-975f-42ce-b1be-ac8a6929bc4d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.666500] env[63021]: DEBUG nova.network.neutron [req-15e7829d-b603-40c4-b0fa-ebc9db5026b0 req-c9acd85a-5b4a-46d2-86ce-bbd4f1e435f0 service nova] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Refreshing network info cache for port eb6c1950-a21a-4873-9c10-53065bf84e01 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 909.680229] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6007c032-6e46-4c78-bbca-672f485c4984 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.682823] env[63021]: DEBUG nova.compute.manager [None req-95b6e88d-8134-48e8-a7c6-826dc81538bf tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 909.683074] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-95b6e88d-8134-48e8-a7c6-826dc81538bf tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 909.692216] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b17d999-b8df-490e-88df-e4f823014c72 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.694682] env[63021]: DEBUG oslo_vmware.api [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Task: {'id': task-1293916, 'name': ReconfigVM_Task, 'duration_secs': 0.389384} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.703165] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Reconfigured VM instance instance-0000004f to attach disk [datastore1] ed26dff1-fed0-4baf-ad41-d14850254aec/ed26dff1-fed0-4baf-ad41-d14850254aec.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 909.704385] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb8a5d1d-bec9-4830-b36d-432c88f95588 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 909.712464] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f61380e7-ab5e-470c-bdfc-e17da4a54368 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.714280] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-260d045a-f006-4fe3-a91c-38bbd7bf4b26 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.716067] env[63021]: DEBUG oslo_vmware.api [None req-64b4d8c9-e71b-42cc-8354-aabb9a08ff12 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293917, 'name': PowerOffVM_Task, 'duration_secs': 0.226408} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.724502] env[63021]: DEBUG oslo_concurrency.lockutils [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.730095] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-64b4d8c9-e71b-42cc-8354-aabb9a08ff12 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 909.735314] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-64b4d8c9-e71b-42cc-8354-aabb9a08ff12 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Reconfiguring VM instance instance-0000004c to detach disk 2001 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 909.737266] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-95b6e88d-8134-48e8-a7c6-826dc81538bf tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 909.737266] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293919, 'name': CreateVM_Task, 'duration_secs': 0.318873} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.737266] env[63021]: DEBUG oslo_vmware.api [None req-92ae3262-30a3-4d02-a8d7-c51a4282461f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1293918, 'name': PowerOffVM_Task, 'duration_secs': 0.236212} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.737266] env[63021]: DEBUG oslo_vmware.api [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1293915, 'name': CreateSnapshot_Task, 'duration_secs': 0.773173} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.741019] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f98a3d0e-3ee9-4444-be47-9afca5f95d69 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.753219] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2c54b6e3-9c84-4e8d-afb2-c8e2145a3034 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.756247] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 909.756610] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-92ae3262-30a3-4d02-a8d7-c51a4282461f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 909.756752] env[63021]: DEBUG nova.compute.manager [None req-92ae3262-30a3-4d02-a8d7-c51a4282461f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 909.757041] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Created Snapshot of the VM instance {{(pid=63021) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 909.760075] env[63021]: DEBUG oslo_concurrency.lockutils [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02de5fc5-60f6-49e2-ae03-e14c840dae1b" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.760273] env[63021]: DEBUG oslo_concurrency.lockutils [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02de5fc5-60f6-49e2-ae03-e14c840dae1b" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.760617] env[63021]: DEBUG oslo_concurrency.lockutils [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02de5fc5-60f6-49e2-ae03-e14c840dae1b" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 909.761376] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d473b1f1-4692-4a37-8a16-cae07252a8de {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.764350] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c876e26-24f3-4b6c-ba59-596003d8f03a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.767932] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b7eaad1-439a-4d4d-9b0a-c53cc1033490 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.771665] env[63021]: DEBUG oslo_vmware.api [None req-cb8a5d1d-bec9-4830-b36d-432c88f95588 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for the task: (returnval){ [ 909.771665] env[63021]: value = "task-1293921" [ 909.771665] env[63021]: _type = "Task" [ 909.771665] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.771914] env[63021]: DEBUG oslo_vmware.api [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Waiting for the task: (returnval){ [ 909.771914] env[63021]: value = "task-1293920" [ 909.771914] env[63021]: _type = "Task" [ 909.771914] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.772807] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a762d4b9-851b-4c36-a89d-96b263bf3f34 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.799548] env[63021]: DEBUG nova.compute.provider_tree [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 909.803083] env[63021]: DEBUG oslo_vmware.api [None req-64b4d8c9-e71b-42cc-8354-aabb9a08ff12 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 909.803083] env[63021]: value = "task-1293922" [ 909.803083] env[63021]: _type = "Task" [ 909.803083] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.803492] env[63021]: DEBUG oslo_vmware.api [None req-95b6e88d-8134-48e8-a7c6-826dc81538bf tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for the task: (returnval){ [ 909.803492] env[63021]: value = "task-1293923" [ 909.803492] env[63021]: _type = "Task" [ 909.803492] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.807304] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 909.807304] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52e6503c-c5c0-dc8e-4491-f1f5dfcbf213" [ 909.807304] env[63021]: _type = "Task" [ 909.807304] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.824289] env[63021]: DEBUG oslo_vmware.api [None req-cb8a5d1d-bec9-4830-b36d-432c88f95588 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293921, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.840897] env[63021]: DEBUG oslo_vmware.api [None req-95b6e88d-8134-48e8-a7c6-826dc81538bf tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293923, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.841236] env[63021]: DEBUG oslo_vmware.api [None req-64b4d8c9-e71b-42cc-8354-aabb9a08ff12 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293922, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.841758] env[63021]: DEBUG oslo_vmware.api [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Task: {'id': task-1293920, 'name': Rename_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.842051] env[63021]: DEBUG oslo_concurrency.lockutils [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02de5fc5-60f6-49e2-ae03-e14c840dae1b" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.842299] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Processing image 02de5fc5-60f6-49e2-ae03-e14c840dae1b {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 909.842573] env[63021]: DEBUG oslo_concurrency.lockutils [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02de5fc5-60f6-49e2-ae03-e14c840dae1b/02de5fc5-60f6-49e2-ae03-e14c840dae1b.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.842659] env[63021]: DEBUG oslo_concurrency.lockutils [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02de5fc5-60f6-49e2-ae03-e14c840dae1b/02de5fc5-60f6-49e2-ae03-e14c840dae1b.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.842836] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 909.843150] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a9e812cd-eefd-4138-890b-f7cea7a4bcdb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.864974] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 909.865263] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 909.866021] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11494302-effa-4015-91c1-1996a90c1480 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.869891] env[63021]: DEBUG nova.network.neutron [None req-65a46cd5-0e5c-4e9c-bb88-b7b7e09d89a7 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Updating instance_info_cache with network_info: [{"id": "1019be13-e855-4eab-8201-a2ae2b22476c", "address": "fa:16:3e:32:70:89", "network": {"id": "708d0dc3-c651-4001-9638-1d3538ff76c9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-983471056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "805af233efa147b3982fb794cd9e6f73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1019be13-e8", "ovs_interfaceid": "1019be13-e855-4eab-8201-a2ae2b22476c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.875907] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 909.875907] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]521f546e-e3c0-e85e-e03a-c7e3aab710c7" [ 909.875907] env[63021]: _type = "Task" [ 909.875907] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.887479] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]521f546e-e3c0-e85e-e03a-c7e3aab710c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.184228] env[63021]: DEBUG oslo_concurrency.lockutils [None req-473cf5dd-c87b-43a2-b8eb-5efd58a3afb2 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Acquiring lock "7c6d7839-616d-41f5-a909-f8c626477a5e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.184475] env[63021]: DEBUG oslo_concurrency.lockutils [None req-473cf5dd-c87b-43a2-b8eb-5efd58a3afb2 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Lock "7c6d7839-616d-41f5-a909-f8c626477a5e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.184711] env[63021]: DEBUG oslo_concurrency.lockutils [None req-473cf5dd-c87b-43a2-b8eb-5efd58a3afb2 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Acquiring lock "7c6d7839-616d-41f5-a909-f8c626477a5e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.184889] env[63021]: DEBUG oslo_concurrency.lockutils [None req-473cf5dd-c87b-43a2-b8eb-5efd58a3afb2 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Lock "7c6d7839-616d-41f5-a909-f8c626477a5e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.185361] env[63021]: DEBUG oslo_concurrency.lockutils [None req-473cf5dd-c87b-43a2-b8eb-5efd58a3afb2 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Lock "7c6d7839-616d-41f5-a909-f8c626477a5e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.191022] env[63021]: INFO nova.compute.manager [None req-473cf5dd-c87b-43a2-b8eb-5efd58a3afb2 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Terminating instance [ 910.192196] env[63021]: DEBUG nova.compute.manager [None req-473cf5dd-c87b-43a2-b8eb-5efd58a3afb2 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 910.192969] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-473cf5dd-c87b-43a2-b8eb-5efd58a3afb2 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 910.194235] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7877e7a9-d081-4bf5-8f47-f96d62d654e3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.200040] env[63021]: DEBUG nova.virt.hardware [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 910.200188] env[63021]: DEBUG nova.virt.hardware [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 910.200835] env[63021]: DEBUG nova.virt.hardware [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 910.201144] env[63021]: DEBUG nova.virt.hardware [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 910.201398] env[63021]: DEBUG nova.virt.hardware [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 910.201583] env[63021]: DEBUG nova.virt.hardware [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 910.201800] env[63021]: DEBUG nova.virt.hardware [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 910.201962] env[63021]: DEBUG nova.virt.hardware [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 910.202149] env[63021]: DEBUG nova.virt.hardware [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 910.202318] env[63021]: DEBUG nova.virt.hardware [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 910.202490] env[63021]: DEBUG nova.virt.hardware [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 910.203282] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4584cba-0ba9-4080-b577-3ea681fa0d8a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.213188] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f28c1d8-3cc9-48ea-8336-8d543de09371 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.216790] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-473cf5dd-c87b-43a2-b8eb-5efd58a3afb2 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 910.217028] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fa0c5f8f-c280-4664-b149-fb197a72cbc9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.229725] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Instance VIF info [] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 910.235466] env[63021]: DEBUG oslo.service.loopingcall [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 910.238860] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 910.239193] env[63021]: DEBUG oslo_vmware.api [None req-473cf5dd-c87b-43a2-b8eb-5efd58a3afb2 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Waiting for the task: (returnval){ [ 910.239193] env[63021]: value = "task-1293924" [ 910.239193] env[63021]: _type = "Task" [ 910.239193] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.239387] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3162931d-8d1f-4544-b4b7-77ae00ffb0bd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.261505] env[63021]: DEBUG oslo_vmware.api [None req-473cf5dd-c87b-43a2-b8eb-5efd58a3afb2 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293924, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.262617] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 910.262617] env[63021]: value = "task-1293925" [ 910.262617] env[63021]: _type = "Task" [ 910.262617] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.270259] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293925, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.287246] env[63021]: DEBUG oslo_vmware.api [None req-cb8a5d1d-bec9-4830-b36d-432c88f95588 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293921, 'name': PowerOffVM_Task, 'duration_secs': 0.204901} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.290367] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb8a5d1d-bec9-4830-b36d-432c88f95588 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 910.290553] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-cb8a5d1d-bec9-4830-b36d-432c88f95588 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 910.291125] env[63021]: DEBUG oslo_vmware.api [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Task: {'id': task-1293920, 'name': Rename_Task, 'duration_secs': 0.229833} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.293475] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dc3e6b65-81ac-481a-8c55-4a0c605b9500 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.294987] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 910.295243] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7df0483d-91ca-4938-b176-c42fe7d53dc4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.301931] env[63021]: DEBUG oslo_vmware.api [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Waiting for the task: (returnval){ [ 910.301931] env[63021]: value = "task-1293926" [ 910.301931] env[63021]: _type = "Task" [ 910.301931] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.308804] env[63021]: DEBUG nova.scheduler.client.report [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 910.312859] env[63021]: DEBUG oslo_concurrency.lockutils [None req-92ae3262-30a3-4d02-a8d7-c51a4282461f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "df856146-61ba-4de7-a5fd-d75263927163" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.156s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.320686] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Creating linked-clone VM from snapshot {{(pid=63021) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 910.332937] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-84a44278-5ef1-43b3-b61d-c80c02c1e347 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.336338] env[63021]: DEBUG oslo_vmware.api [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Task: {'id': task-1293926, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.342353] env[63021]: DEBUG oslo_vmware.api [None req-64b4d8c9-e71b-42cc-8354-aabb9a08ff12 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293922, 'name': ReconfigVM_Task, 'duration_secs': 0.277814} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.346412] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-64b4d8c9-e71b-42cc-8354-aabb9a08ff12 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Reconfigured VM instance instance-0000004c to detach disk 2001 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 910.346551] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-64b4d8c9-e71b-42cc-8354-aabb9a08ff12 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 910.346832] env[63021]: DEBUG oslo_vmware.api [None req-95b6e88d-8134-48e8-a7c6-826dc81538bf tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293923, 'name': PowerOffVM_Task, 'duration_secs': 0.225051} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.347111] env[63021]: DEBUG oslo_vmware.api [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Waiting for the task: (returnval){ [ 910.347111] env[63021]: value = "task-1293928" [ 910.347111] env[63021]: _type = "Task" [ 910.347111] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.347641] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-42947219-8ff1-4e91-a202-8cbe30aaf324 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.349446] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-95b6e88d-8134-48e8-a7c6-826dc81538bf tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 910.349657] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-95b6e88d-8134-48e8-a7c6-826dc81538bf tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 910.350015] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3d48b746-2a6a-4b6b-b3e0-cb12afff0df1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.360587] env[63021]: DEBUG oslo_vmware.api [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1293928, 'name': CloneVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.362091] env[63021]: DEBUG oslo_vmware.api [None req-64b4d8c9-e71b-42cc-8354-aabb9a08ff12 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 910.362091] env[63021]: value = "task-1293929" [ 910.362091] env[63021]: _type = "Task" [ 910.362091] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.372955] env[63021]: DEBUG oslo_vmware.api [None req-64b4d8c9-e71b-42cc-8354-aabb9a08ff12 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293929, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.373323] env[63021]: DEBUG oslo_concurrency.lockutils [None req-65a46cd5-0e5c-4e9c-bb88-b7b7e09d89a7 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Releasing lock "refresh_cache-71909213-258d-45f5-9a3e-e473e8ce8aa2" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.373560] env[63021]: DEBUG nova.objects.instance [None req-65a46cd5-0e5c-4e9c-bb88-b7b7e09d89a7 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lazy-loading 'migration_context' on Instance uuid 71909213-258d-45f5-9a3e-e473e8ce8aa2 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 910.385313] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-cb8a5d1d-bec9-4830-b36d-432c88f95588 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 910.385590] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-cb8a5d1d-bec9-4830-b36d-432c88f95588 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 910.385784] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb8a5d1d-bec9-4830-b36d-432c88f95588 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Deleting the datastore file [datastore1] 44dadf8e-a727-448f-887d-6408f7ebfbc6 {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 910.389362] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e9dca7d1-6901-4e17-bb66-ffc5b718c693 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.391955] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Preparing fetch location {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 910.392240] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Fetch image to [datastore1] OSTACK_IMG_00c60d10-c465-4c17-b188-9861d55eeda2/OSTACK_IMG_00c60d10-c465-4c17-b188-9861d55eeda2.vmdk {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 910.392435] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Downloading stream optimized image 02de5fc5-60f6-49e2-ae03-e14c840dae1b to [datastore1] OSTACK_IMG_00c60d10-c465-4c17-b188-9861d55eeda2/OSTACK_IMG_00c60d10-c465-4c17-b188-9861d55eeda2.vmdk on the data store datastore1 as vApp {{(pid=63021) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 910.392605] env[63021]: DEBUG nova.virt.vmwareapi.images [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Downloading image file data 02de5fc5-60f6-49e2-ae03-e14c840dae1b to the ESX as VM named 'OSTACK_IMG_00c60d10-c465-4c17-b188-9861d55eeda2' {{(pid=63021) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 910.404020] env[63021]: DEBUG oslo_vmware.api [None req-cb8a5d1d-bec9-4830-b36d-432c88f95588 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for the task: (returnval){ [ 910.404020] env[63021]: value = "task-1293931" [ 910.404020] env[63021]: _type = "Task" [ 910.404020] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.414863] env[63021]: DEBUG oslo_vmware.api [None req-cb8a5d1d-bec9-4830-b36d-432c88f95588 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293931, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.443608] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-95b6e88d-8134-48e8-a7c6-826dc81538bf tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 910.444443] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-95b6e88d-8134-48e8-a7c6-826dc81538bf tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 910.444443] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-95b6e88d-8134-48e8-a7c6-826dc81538bf tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Deleting the datastore file [datastore1] 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 910.444443] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cb27ec3f-fe53-4115-9938-5eb7772ac116 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.450474] env[63021]: DEBUG oslo_vmware.api [None req-95b6e88d-8134-48e8-a7c6-826dc81538bf tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for the task: (returnval){ [ 910.450474] env[63021]: value = "task-1293932" [ 910.450474] env[63021]: _type = "Task" [ 910.450474] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.463086] env[63021]: DEBUG oslo_vmware.api [None req-95b6e88d-8134-48e8-a7c6-826dc81538bf tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293932, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.488090] env[63021]: DEBUG oslo_vmware.rw_handles [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 910.488090] env[63021]: value = "resgroup-9" [ 910.488090] env[63021]: _type = "ResourcePool" [ 910.488090] env[63021]: }. {{(pid=63021) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 910.488482] env[63021]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-97894db9-2506-48aa-8711-2a935b6c2f8e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.510107] env[63021]: DEBUG oslo_vmware.rw_handles [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lease: (returnval){ [ 910.510107] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]525c3782-5adb-be1d-5ea7-4f606cddc5ed" [ 910.510107] env[63021]: _type = "HttpNfcLease" [ 910.510107] env[63021]: } obtained for vApp import into resource pool (val){ [ 910.510107] env[63021]: value = "resgroup-9" [ 910.510107] env[63021]: _type = "ResourcePool" [ 910.510107] env[63021]: }. {{(pid=63021) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 910.510480] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the lease: (returnval){ [ 910.510480] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]525c3782-5adb-be1d-5ea7-4f606cddc5ed" [ 910.510480] env[63021]: _type = "HttpNfcLease" [ 910.510480] env[63021]: } to be ready. {{(pid=63021) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 910.514797] env[63021]: DEBUG nova.network.neutron [req-15e7829d-b603-40c4-b0fa-ebc9db5026b0 req-c9acd85a-5b4a-46d2-86ce-bbd4f1e435f0 service nova] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Updated VIF entry in instance network info cache for port eb6c1950-a21a-4873-9c10-53065bf84e01. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 910.515200] env[63021]: DEBUG nova.network.neutron [req-15e7829d-b603-40c4-b0fa-ebc9db5026b0 req-c9acd85a-5b4a-46d2-86ce-bbd4f1e435f0 service nova] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Updating instance_info_cache with network_info: [{"id": "eb6c1950-a21a-4873-9c10-53065bf84e01", "address": "fa:16:3e:b9:50:07", "network": {"id": "265e2e2e-6ec6-4d4c-8b4c-0a2e8baa6bdb", "bridge": "br-int", "label": "tempest-ImagesTestJSON-242691215-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7331caed42034dfc81c79cfabbe0829e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bfae3ef8-cae7-455d-8632-ba93e1671625", "external-id": "cl2-zone-841", "segmentation_id": 841, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb6c1950-a2", "ovs_interfaceid": "eb6c1950-a21a-4873-9c10-53065bf84e01", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.517705] env[63021]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 910.517705] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]525c3782-5adb-be1d-5ea7-4f606cddc5ed" [ 910.517705] env[63021]: _type = "HttpNfcLease" [ 910.517705] env[63021]: } is initializing. {{(pid=63021) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 910.762357] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3be93392-7165-4514-87ff-a9b7ec483ec6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "c328d2f7-3398-4f25-b11c-f464be7af8a1" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.762771] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3be93392-7165-4514-87ff-a9b7ec483ec6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "c328d2f7-3398-4f25-b11c-f464be7af8a1" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.762915] env[63021]: DEBUG nova.compute.manager [None req-3be93392-7165-4514-87ff-a9b7ec483ec6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 910.763269] env[63021]: DEBUG oslo_vmware.api [None req-473cf5dd-c87b-43a2-b8eb-5efd58a3afb2 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293924, 'name': PowerOffVM_Task, 'duration_secs': 0.218663} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.764165] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7b639ee-343f-4986-8dc0-801cf0fac6ed {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.768094] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-473cf5dd-c87b-43a2-b8eb-5efd58a3afb2 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 910.768380] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-473cf5dd-c87b-43a2-b8eb-5efd58a3afb2 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 910.772144] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-876b430b-3100-47a6-a4dc-ecce82a4da42 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.778813] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293925, 'name': CreateVM_Task, 'duration_secs': 0.48477} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.780218] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 910.780552] env[63021]: DEBUG nova.compute.manager [None req-3be93392-7165-4514-87ff-a9b7ec483ec6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63021) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 910.781139] env[63021]: DEBUG nova.objects.instance [None req-3be93392-7165-4514-87ff-a9b7ec483ec6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lazy-loading 'flavor' on Instance uuid c328d2f7-3398-4f25-b11c-f464be7af8a1 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 910.782922] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.783036] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.783338] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 910.783946] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ed0f142-600c-41a5-99ea-a61a82962ae0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.788435] env[63021]: DEBUG oslo_vmware.api [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Waiting for the task: (returnval){ [ 910.788435] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5218f62f-ddc7-faf3-e666-2e2491ab6d19" [ 910.788435] env[63021]: _type = "Task" [ 910.788435] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.799713] env[63021]: DEBUG oslo_vmware.api [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5218f62f-ddc7-faf3-e666-2e2491ab6d19, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.811145] env[63021]: DEBUG oslo_vmware.api [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Task: {'id': task-1293926, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.815159] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.707s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.817308] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.474s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.818825] env[63021]: INFO nova.compute.claims [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 910.837869] env[63021]: INFO nova.scheduler.client.report [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Deleted allocations for instance 36ec45ec-761b-4d62-a74f-e4d9a840ada0 [ 910.860850] env[63021]: DEBUG oslo_vmware.api [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1293928, 'name': CloneVM_Task} progress is 94%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.872030] env[63021]: DEBUG oslo_vmware.api [None req-64b4d8c9-e71b-42cc-8354-aabb9a08ff12 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293929, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.876966] env[63021]: DEBUG nova.objects.base [None req-65a46cd5-0e5c-4e9c-bb88-b7b7e09d89a7 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Object Instance<71909213-258d-45f5-9a3e-e473e8ce8aa2> lazy-loaded attributes: info_cache,migration_context {{(pid=63021) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 910.878218] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98597b93-7871-4e8e-a68b-8e9d81c30cbe {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.900860] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec625c04-8fbc-465a-8fef-479095d371e2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.903956] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-473cf5dd-c87b-43a2-b8eb-5efd58a3afb2 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 910.904241] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-473cf5dd-c87b-43a2-b8eb-5efd58a3afb2 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 910.904448] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-473cf5dd-c87b-43a2-b8eb-5efd58a3afb2 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Deleting the datastore file [datastore1] 7c6d7839-616d-41f5-a909-f8c626477a5e {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 910.909417] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f6761824-a3e8-4b26-865d-5b1597a99ebe {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.913888] env[63021]: DEBUG oslo_vmware.api [None req-65a46cd5-0e5c-4e9c-bb88-b7b7e09d89a7 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 910.913888] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]521a33a8-a745-f843-8e84-c1c2360bc69b" [ 910.913888] env[63021]: _type = "Task" [ 910.913888] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.917261] env[63021]: DEBUG oslo_vmware.api [None req-cb8a5d1d-bec9-4830-b36d-432c88f95588 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293931, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.245588} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.921235] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb8a5d1d-bec9-4830-b36d-432c88f95588 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 910.921437] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-cb8a5d1d-bec9-4830-b36d-432c88f95588 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 910.921615] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-cb8a5d1d-bec9-4830-b36d-432c88f95588 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 910.921779] env[63021]: INFO nova.compute.manager [None req-cb8a5d1d-bec9-4830-b36d-432c88f95588 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Took 1.28 seconds to destroy the instance on the hypervisor. [ 910.922016] env[63021]: DEBUG oslo.service.loopingcall [None req-cb8a5d1d-bec9-4830-b36d-432c88f95588 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 910.922291] env[63021]: DEBUG oslo_vmware.api [None req-473cf5dd-c87b-43a2-b8eb-5efd58a3afb2 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Waiting for the task: (returnval){ [ 910.922291] env[63021]: value = "task-1293935" [ 910.922291] env[63021]: _type = "Task" [ 910.922291] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.922539] env[63021]: DEBUG nova.compute.manager [-] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 910.922608] env[63021]: DEBUG nova.network.neutron [-] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 910.930973] env[63021]: DEBUG oslo_vmware.api [None req-65a46cd5-0e5c-4e9c-bb88-b7b7e09d89a7 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]521a33a8-a745-f843-8e84-c1c2360bc69b, 'name': SearchDatastore_Task, 'duration_secs': 0.010041} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.931640] env[63021]: DEBUG oslo_concurrency.lockutils [None req-65a46cd5-0e5c-4e9c-bb88-b7b7e09d89a7 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.935622] env[63021]: DEBUG oslo_vmware.api [None req-473cf5dd-c87b-43a2-b8eb-5efd58a3afb2 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293935, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.960540] env[63021]: DEBUG oslo_vmware.api [None req-95b6e88d-8134-48e8-a7c6-826dc81538bf tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Task: {'id': task-1293932, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.232491} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.960840] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-95b6e88d-8134-48e8-a7c6-826dc81538bf tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 910.961035] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-95b6e88d-8134-48e8-a7c6-826dc81538bf tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 910.961221] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-95b6e88d-8134-48e8-a7c6-826dc81538bf tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 910.961401] env[63021]: INFO nova.compute.manager [None req-95b6e88d-8134-48e8-a7c6-826dc81538bf tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Took 1.28 seconds to destroy the instance on the hypervisor. [ 910.961641] env[63021]: DEBUG oslo.service.loopingcall [None req-95b6e88d-8134-48e8-a7c6-826dc81538bf tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 910.961838] env[63021]: DEBUG nova.compute.manager [-] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 910.961927] env[63021]: DEBUG nova.network.neutron [-] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 911.019672] env[63021]: DEBUG oslo_concurrency.lockutils [req-15e7829d-b603-40c4-b0fa-ebc9db5026b0 req-c9acd85a-5b4a-46d2-86ce-bbd4f1e435f0 service nova] Releasing lock "refresh_cache-d63fd465-975f-42ce-b1be-ac8a6929bc4d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.020120] env[63021]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 911.020120] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]525c3782-5adb-be1d-5ea7-4f606cddc5ed" [ 911.020120] env[63021]: _type = "HttpNfcLease" [ 911.020120] env[63021]: } is initializing. {{(pid=63021) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 911.101813] env[63021]: DEBUG nova.objects.instance [None req-96ce6cf7-19a2-4299-affa-85e2a3ac4827 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lazy-loading 'flavor' on Instance uuid df856146-61ba-4de7-a5fd-d75263927163 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 911.287543] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-3be93392-7165-4514-87ff-a9b7ec483ec6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 911.287916] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-870ae3d2-5cb5-4212-a392-b887b5f81dc0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.299088] env[63021]: DEBUG oslo_vmware.api [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5218f62f-ddc7-faf3-e666-2e2491ab6d19, 'name': SearchDatastore_Task, 'duration_secs': 0.011517} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.300441] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.300888] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 911.300969] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.301139] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.301437] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 911.301910] env[63021]: DEBUG oslo_vmware.api [None req-3be93392-7165-4514-87ff-a9b7ec483ec6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 911.301910] env[63021]: value = "task-1293936" [ 911.301910] env[63021]: _type = "Task" [ 911.301910] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.302198] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ce849664-cfc3-455b-b392-a879f567f684 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.315282] env[63021]: DEBUG oslo_vmware.api [None req-3be93392-7165-4514-87ff-a9b7ec483ec6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1293936, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.328176] env[63021]: DEBUG oslo_vmware.api [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Task: {'id': task-1293926, 'name': PowerOnVM_Task, 'duration_secs': 0.64139} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.329609] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 911.329824] env[63021]: INFO nova.compute.manager [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Took 11.38 seconds to spawn the instance on the hypervisor. [ 911.330060] env[63021]: DEBUG nova.compute.manager [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 911.330311] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 911.330472] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 911.334404] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6add88bd-51c8-4112-a9f0-dd06eb3aac47 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.337242] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ccd03a04-ec43-43f1-a075-b30850300ad3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.354129] env[63021]: DEBUG oslo_vmware.api [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Waiting for the task: (returnval){ [ 911.354129] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52e8c0a8-2a9f-1389-4a77-f6e7f3314df5" [ 911.354129] env[63021]: _type = "Task" [ 911.354129] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.354715] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85dbc593-bfaf-49e3-8b4a-5980d50cdd35 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.490s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.356597] env[63021]: DEBUG oslo_concurrency.lockutils [req-227023fa-159b-4e6b-9bd5-4071fb5a5e92 req-dc68025a-81f9-4788-940e-bca9bd14141f service nova] Acquired lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.364426] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ae40072-4909-4c91-bbf9-cc620abb0d26 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.379231] env[63021]: DEBUG oslo_vmware.api [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1293928, 'name': CloneVM_Task} progress is 94%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.380770] env[63021]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 911.381019] env[63021]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=63021) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 911.386040] env[63021]: DEBUG oslo_vmware.api [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52e8c0a8-2a9f-1389-4a77-f6e7f3314df5, 'name': SearchDatastore_Task, 'duration_secs': 0.009504} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.386855] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-01464027-a542-48a9-af50-51a73094f122 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.394042] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-18561ebf-aa9d-4c4c-9509-8fe0d72dad76 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.397465] env[63021]: DEBUG nova.compute.manager [req-fb3ce6c5-f790-4782-9865-c76ffabb9bdb req-0bcee101-a1ea-416d-bd07-e2c431635469 service nova] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Received event network-vif-deleted-fb56a4a1-033a-4205-aced-43f632d2f594 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 911.397725] env[63021]: INFO nova.compute.manager [req-fb3ce6c5-f790-4782-9865-c76ffabb9bdb req-0bcee101-a1ea-416d-bd07-e2c431635469 service nova] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Neutron deleted interface fb56a4a1-033a-4205-aced-43f632d2f594; detaching it from the instance and deleting it from the info cache [ 911.397973] env[63021]: DEBUG nova.network.neutron [req-fb3ce6c5-f790-4782-9865-c76ffabb9bdb req-0bcee101-a1ea-416d-bd07-e2c431635469 service nova] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.399257] env[63021]: DEBUG oslo_vmware.api [None req-64b4d8c9-e71b-42cc-8354-aabb9a08ff12 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293929, 'name': PowerOnVM_Task, 'duration_secs': 0.550786} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.400654] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-64b4d8c9-e71b-42cc-8354-aabb9a08ff12 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 911.400971] env[63021]: DEBUG nova.compute.manager [None req-64b4d8c9-e71b-42cc-8354-aabb9a08ff12 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 911.402479] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2228d9b5-d8d6-46f5-95b1-4fa7327c743a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.407457] env[63021]: DEBUG oslo_vmware.api [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Waiting for the task: (returnval){ [ 911.407457] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52f2e133-70ed-a18e-13d8-436dda4bbc20" [ 911.407457] env[63021]: _type = "Task" [ 911.407457] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.412779] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2e4904a-153e-4e30-bdc1-4f084b8bf1fa {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.440865] env[63021]: DEBUG oslo_vmware.api [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52f2e133-70ed-a18e-13d8-436dda4bbc20, 'name': SearchDatastore_Task, 'duration_secs': 0.017904} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.453512] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.453841] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 4d712cc1-6202-407e-8767-c620a1ce0bba/4d712cc1-6202-407e-8767-c620a1ce0bba.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 911.454171] env[63021]: DEBUG oslo_vmware.api [None req-473cf5dd-c87b-43a2-b8eb-5efd58a3afb2 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Task: {'id': task-1293935, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.22812} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.456223] env[63021]: ERROR root [req-227023fa-159b-4e6b-9bd5-4071fb5a5e92 req-dc68025a-81f9-4788-940e-bca9bd14141f service nova] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler\n response = request(managed_object, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__\n return client.invoke(args, kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke\n result = self.send(soapenv, timeout=timeout)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send\n return self.process_reply(reply.message, None, None)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply\n raise WebFault(fault, replyroot)\n', "suds.WebFault: Server raised fault: 'The object 'vim.VirtualMachine:vm-277478' has already been deleted or has not been completely created'\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api\n return api_method(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 480, in get_object_property\n props = get_object_properties(vim, moref, [property_name],\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 360, in get_object_properties\n retrieve_result = vim.RetrievePropertiesEx(\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler\n raise exceptions.VimFaultException(fault_list, fault_string,\n', "oslo_vmware.exceptions.VimFaultException: The object 'vim.VirtualMachine:vm-277478' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-277478' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-277478'}\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 123, in _call_method\n return self.invoke_api(module, method, self.vim, *args,\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api\n return _invoke_api(module, method, *args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func\n return evt.wait()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait\n result = hub.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch\n return self.greenlet.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner\n idle = self.f(*self.args, **self.kw)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func\n result = f(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api\n raise clazz(str(excep),\n', "oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-277478' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-277478' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-277478'}\n"]: nova.exception.InstanceNotFound: Instance 36ec45ec-761b-4d62-a74f-e4d9a840ada0 could not be found. [ 911.456431] env[63021]: DEBUG oslo_concurrency.lockutils [req-227023fa-159b-4e6b-9bd5-4071fb5a5e92 req-dc68025a-81f9-4788-940e-bca9bd14141f service nova] Releasing lock "36ec45ec-761b-4d62-a74f-e4d9a840ada0" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.456638] env[63021]: DEBUG nova.compute.manager [req-227023fa-159b-4e6b-9bd5-4071fb5a5e92 req-dc68025a-81f9-4788-940e-bca9bd14141f service nova] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Detach interface failed, port_id=d92edf1f-d15d-4f56-9da3-92cbe44b6709, reason: Instance 36ec45ec-761b-4d62-a74f-e4d9a840ada0 could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 911.457219] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2f203fb9-83d9-454c-855d-b77b20b3f6ce {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.459289] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-473cf5dd-c87b-43a2-b8eb-5efd58a3afb2 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 911.459482] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-473cf5dd-c87b-43a2-b8eb-5efd58a3afb2 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 911.459680] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-473cf5dd-c87b-43a2-b8eb-5efd58a3afb2 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 911.459851] env[63021]: INFO nova.compute.manager [None req-473cf5dd-c87b-43a2-b8eb-5efd58a3afb2 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Took 1.27 seconds to destroy the instance on the hypervisor. [ 911.460102] env[63021]: DEBUG oslo.service.loopingcall [None req-473cf5dd-c87b-43a2-b8eb-5efd58a3afb2 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 911.460353] env[63021]: DEBUG nova.compute.manager [-] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 911.460443] env[63021]: DEBUG nova.network.neutron [-] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 911.468032] env[63021]: DEBUG oslo_vmware.api [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Waiting for the task: (returnval){ [ 911.468032] env[63021]: value = "task-1293937" [ 911.468032] env[63021]: _type = "Task" [ 911.468032] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.475993] env[63021]: DEBUG oslo_vmware.api [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293937, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.519258] env[63021]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 911.519258] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]525c3782-5adb-be1d-5ea7-4f606cddc5ed" [ 911.519258] env[63021]: _type = "HttpNfcLease" [ 911.519258] env[63021]: } is ready. {{(pid=63021) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 911.519556] env[63021]: DEBUG oslo_vmware.rw_handles [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 911.519556] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]525c3782-5adb-be1d-5ea7-4f606cddc5ed" [ 911.519556] env[63021]: _type = "HttpNfcLease" [ 911.519556] env[63021]: }. {{(pid=63021) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 911.520326] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2def6bfc-48d4-406c-a3f3-460bdb9a5be0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.530116] env[63021]: DEBUG oslo_vmware.rw_handles [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52010ce0-7aed-0ada-0b25-c956ac1ff706/disk-0.vmdk from lease info. {{(pid=63021) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 911.531137] env[63021]: DEBUG oslo_vmware.rw_handles [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52010ce0-7aed-0ada-0b25-c956ac1ff706/disk-0.vmdk. {{(pid=63021) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 911.595226] env[63021]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-cf143e29-cf8a-4a1c-b9f5-c91d1646eb72 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.608877] env[63021]: DEBUG oslo_concurrency.lockutils [None req-96ce6cf7-19a2-4299-affa-85e2a3ac4827 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquiring lock "refresh_cache-df856146-61ba-4de7-a5fd-d75263927163" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.608877] env[63021]: DEBUG oslo_concurrency.lockutils [None req-96ce6cf7-19a2-4299-affa-85e2a3ac4827 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquired lock "refresh_cache-df856146-61ba-4de7-a5fd-d75263927163" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.608877] env[63021]: DEBUG nova.network.neutron [None req-96ce6cf7-19a2-4299-affa-85e2a3ac4827 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 911.608877] env[63021]: DEBUG nova.objects.instance [None req-96ce6cf7-19a2-4299-affa-85e2a3ac4827 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lazy-loading 'info_cache' on Instance uuid df856146-61ba-4de7-a5fd-d75263927163 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 911.814875] env[63021]: DEBUG oslo_vmware.api [None req-3be93392-7165-4514-87ff-a9b7ec483ec6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1293936, 'name': PowerOffVM_Task, 'duration_secs': 0.269753} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.815335] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-3be93392-7165-4514-87ff-a9b7ec483ec6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 911.815393] env[63021]: DEBUG nova.compute.manager [None req-3be93392-7165-4514-87ff-a9b7ec483ec6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 911.816198] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fc9a10f-ade7-40e3-b60f-d5bea4263f39 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.865648] env[63021]: INFO nova.compute.manager [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Took 41.15 seconds to build instance. [ 911.869426] env[63021]: DEBUG nova.network.neutron [-] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.886072] env[63021]: DEBUG nova.network.neutron [-] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.887485] env[63021]: DEBUG oslo_vmware.api [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1293928, 'name': CloneVM_Task, 'duration_secs': 1.454537} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.888601] env[63021]: INFO nova.virt.vmwareapi.vmops [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Created linked-clone VM from snapshot [ 911.889122] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3dd91b9-7b23-4648-8386-181b32cba638 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.903255] env[63021]: DEBUG nova.virt.vmwareapi.images [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Uploading image d9df5b42-2840-4afc-82fa-268aed408575 {{(pid=63021) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 911.905625] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0beb391f-252c-4419-ac46-097b23e279a8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.917706] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-125ab548-8cc6-451a-a2ef-63456e7b6fd1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.945040] env[63021]: DEBUG oslo_vmware.rw_handles [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 911.945040] env[63021]: value = "vm-277547" [ 911.945040] env[63021]: _type = "VirtualMachine" [ 911.945040] env[63021]: }. {{(pid=63021) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 911.945479] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-e86623fd-ab30-4d2c-8f89-f6242480d6d8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.959171] env[63021]: DEBUG nova.compute.manager [req-fb3ce6c5-f790-4782-9865-c76ffabb9bdb req-0bcee101-a1ea-416d-bd07-e2c431635469 service nova] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Detach interface failed, port_id=fb56a4a1-033a-4205-aced-43f632d2f594, reason: Instance 44dadf8e-a727-448f-887d-6408f7ebfbc6 could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 911.969096] env[63021]: DEBUG oslo_vmware.rw_handles [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lease: (returnval){ [ 911.969096] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]523e7e29-458e-92b6-7ce9-d9e640d1d1d1" [ 911.969096] env[63021]: _type = "HttpNfcLease" [ 911.969096] env[63021]: } obtained for exporting VM: (result){ [ 911.969096] env[63021]: value = "vm-277547" [ 911.969096] env[63021]: _type = "VirtualMachine" [ 911.969096] env[63021]: }. {{(pid=63021) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 911.969694] env[63021]: DEBUG oslo_vmware.api [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Waiting for the lease: (returnval){ [ 911.969694] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]523e7e29-458e-92b6-7ce9-d9e640d1d1d1" [ 911.969694] env[63021]: _type = "HttpNfcLease" [ 911.969694] env[63021]: } to be ready. {{(pid=63021) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 911.987398] env[63021]: DEBUG oslo_vmware.api [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293937, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.989288] env[63021]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 911.989288] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]523e7e29-458e-92b6-7ce9-d9e640d1d1d1" [ 911.989288] env[63021]: _type = "HttpNfcLease" [ 911.989288] env[63021]: } is initializing. {{(pid=63021) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 912.115470] env[63021]: DEBUG nova.objects.base [None req-96ce6cf7-19a2-4299-affa-85e2a3ac4827 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=63021) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 912.206024] env[63021]: DEBUG nova.network.neutron [-] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.215572] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a809eaa5-a63b-45da-abc2-116713fda49d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.224522] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4fe5c3c-18f1-4835-83d2-cdf92d8d94a7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.260398] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e923066d-974e-49e7-8337-cf7b2ab0305d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.269330] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb234216-b08a-4341-8ad5-d8a64f610ea5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.285104] env[63021]: DEBUG nova.compute.provider_tree [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 912.328109] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3be93392-7165-4514-87ff-a9b7ec483ec6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "c328d2f7-3398-4f25-b11c-f464be7af8a1" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.565s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.375651] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5ce131bf-229b-494d-a710-f7b81db2ddc1 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Lock "ed26dff1-fed0-4baf-ad41-d14850254aec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.816s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.376287] env[63021]: INFO nova.compute.manager [-] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Took 1.45 seconds to deallocate network for instance. [ 912.389766] env[63021]: INFO nova.compute.manager [-] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Took 1.43 seconds to deallocate network for instance. [ 912.481719] env[63021]: DEBUG oslo_vmware.api [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293937, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.785289} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.482412] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 4d712cc1-6202-407e-8767-c620a1ce0bba/4d712cc1-6202-407e-8767-c620a1ce0bba.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 912.482639] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 912.483833] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cc7acefd-6653-40f9-b61b-c2b4405222e5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.487519] env[63021]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 912.487519] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]523e7e29-458e-92b6-7ce9-d9e640d1d1d1" [ 912.487519] env[63021]: _type = "HttpNfcLease" [ 912.487519] env[63021]: } is ready. {{(pid=63021) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 912.489045] env[63021]: DEBUG oslo_vmware.rw_handles [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 912.489045] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]523e7e29-458e-92b6-7ce9-d9e640d1d1d1" [ 912.489045] env[63021]: _type = "HttpNfcLease" [ 912.489045] env[63021]: }. {{(pid=63021) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 912.489784] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5442ae71-5df0-4197-9cd4-4372887a6b8b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.494040] env[63021]: DEBUG oslo_vmware.api [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Waiting for the task: (returnval){ [ 912.494040] env[63021]: value = "task-1293939" [ 912.494040] env[63021]: _type = "Task" [ 912.494040] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.500395] env[63021]: DEBUG oslo_vmware.rw_handles [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a434df-1d61-e144-a3f8-ffe1b46c8c86/disk-0.vmdk from lease info. {{(pid=63021) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 912.500640] env[63021]: DEBUG oslo_vmware.rw_handles [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a434df-1d61-e144-a3f8-ffe1b46c8c86/disk-0.vmdk for reading. {{(pid=63021) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 912.573517] env[63021]: DEBUG oslo_vmware.api [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293939, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.633427] env[63021]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-01d8e358-401d-4345-908c-763d4f5baceb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.706979] env[63021]: INFO nova.compute.manager [-] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Took 1.25 seconds to deallocate network for instance. [ 912.791670] env[63021]: DEBUG nova.scheduler.client.report [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 912.888147] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cb8a5d1d-bec9-4830-b36d-432c88f95588 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.888147] env[63021]: DEBUG oslo_vmware.rw_handles [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Completed reading data from the image iterator. {{(pid=63021) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 912.888147] env[63021]: DEBUG oslo_vmware.rw_handles [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52010ce0-7aed-0ada-0b25-c956ac1ff706/disk-0.vmdk. {{(pid=63021) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 912.890510] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55344fa6-f00a-4f36-8941-53381c2407dd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.899119] env[63021]: DEBUG oslo_concurrency.lockutils [None req-95b6e88d-8134-48e8-a7c6-826dc81538bf tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.901280] env[63021]: DEBUG oslo_vmware.rw_handles [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52010ce0-7aed-0ada-0b25-c956ac1ff706/disk-0.vmdk is in state: ready. {{(pid=63021) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 912.901280] env[63021]: DEBUG oslo_vmware.rw_handles [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52010ce0-7aed-0ada-0b25-c956ac1ff706/disk-0.vmdk. {{(pid=63021) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 912.901509] env[63021]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-82042f3e-075d-4def-a46c-98b9e48ae59f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.011436] env[63021]: DEBUG oslo_vmware.api [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293939, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073339} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.011965] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 913.013372] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e73b51d8-c8ee-442f-9f55-63df211c8851 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.041052] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] 4d712cc1-6202-407e-8767-c620a1ce0bba/4d712cc1-6202-407e-8767-c620a1ce0bba.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 913.041596] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0d5227ec-9cbc-46c5-b9c4-6a07a768f50a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.060269] env[63021]: DEBUG nova.network.neutron [None req-96ce6cf7-19a2-4299-affa-85e2a3ac4827 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Updating instance_info_cache with network_info: [{"id": "a43dcfd2-397e-44fe-bd1c-b6116c25a28a", "address": "fa:16:3e:c8:73:42", "network": {"id": "7258df3e-b1eb-4da6-ab52-5dadd2f87b51", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1084458843-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.144", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5661c44b534b4a7e9452e914fc80d66e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4f04accd-4a9f-41a1-b327-c96095e1ade4", "external-id": "nsx-vlan-transportzone-947", "segmentation_id": 947, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa43dcfd2-39", "ovs_interfaceid": "a43dcfd2-397e-44fe-bd1c-b6116c25a28a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.069756] env[63021]: DEBUG oslo_vmware.api [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Waiting for the task: (returnval){ [ 913.069756] env[63021]: value = "task-1293940" [ 913.069756] env[63021]: _type = "Task" [ 913.069756] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.084046] env[63021]: DEBUG oslo_vmware.api [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293940, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.084046] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquiring lock "1b4871cc-3dac-4683-8fce-93c4d9e4407d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.084046] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Lock "1b4871cc-3dac-4683-8fce-93c4d9e4407d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.084046] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquiring lock "1b4871cc-3dac-4683-8fce-93c4d9e4407d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.084046] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Lock "1b4871cc-3dac-4683-8fce-93c4d9e4407d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.084046] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Lock "1b4871cc-3dac-4683-8fce-93c4d9e4407d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.086713] env[63021]: INFO nova.compute.manager [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Terminating instance [ 913.090343] env[63021]: DEBUG nova.compute.manager [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 913.090343] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 913.091108] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f176f8f-2361-4f64-bf9c-29a47f69802e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.100182] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 913.100488] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d0a5bc8a-5635-48ff-b9c3-f91ea4e3fc8e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.109183] env[63021]: DEBUG oslo_vmware.api [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 913.109183] env[63021]: value = "task-1293941" [ 913.109183] env[63021]: _type = "Task" [ 913.109183] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.119639] env[63021]: DEBUG oslo_vmware.api [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293941, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.157969] env[63021]: DEBUG nova.compute.manager [req-c568f9b6-4edb-430a-bc9a-e8ceede4d0f7 req-0f3fe796-c71c-44a8-a831-0a893a133e16 service nova] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Received event network-changed-2e8a7fc6-825c-45a5-b41c-d28410061aa3 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 913.158367] env[63021]: DEBUG nova.compute.manager [req-c568f9b6-4edb-430a-bc9a-e8ceede4d0f7 req-0f3fe796-c71c-44a8-a831-0a893a133e16 service nova] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Refreshing instance network info cache due to event network-changed-2e8a7fc6-825c-45a5-b41c-d28410061aa3. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 913.158708] env[63021]: DEBUG oslo_concurrency.lockutils [req-c568f9b6-4edb-430a-bc9a-e8ceede4d0f7 req-0f3fe796-c71c-44a8-a831-0a893a133e16 service nova] Acquiring lock "refresh_cache-ed26dff1-fed0-4baf-ad41-d14850254aec" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.159016] env[63021]: DEBUG oslo_concurrency.lockutils [req-c568f9b6-4edb-430a-bc9a-e8ceede4d0f7 req-0f3fe796-c71c-44a8-a831-0a893a133e16 service nova] Acquired lock "refresh_cache-ed26dff1-fed0-4baf-ad41-d14850254aec" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.159294] env[63021]: DEBUG nova.network.neutron [req-c568f9b6-4edb-430a-bc9a-e8ceede4d0f7 req-0f3fe796-c71c-44a8-a831-0a893a133e16 service nova] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Refreshing network info cache for port 2e8a7fc6-825c-45a5-b41c-d28410061aa3 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 913.215722] env[63021]: DEBUG oslo_concurrency.lockutils [None req-473cf5dd-c87b-43a2-b8eb-5efd58a3afb2 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.245436] env[63021]: DEBUG oslo_vmware.rw_handles [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52010ce0-7aed-0ada-0b25-c956ac1ff706/disk-0.vmdk. {{(pid=63021) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 913.245975] env[63021]: INFO nova.virt.vmwareapi.images [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Downloaded image file data 02de5fc5-60f6-49e2-ae03-e14c840dae1b [ 913.247095] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67e87d17-77f3-47be-8d52-8cf0a1f0ae61 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.266873] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bbce45b2-c5ae-494c-8cbf-a116b46ff10a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.303023] env[63021]: INFO nova.virt.vmwareapi.images [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] The imported VM was unregistered [ 913.304430] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Caching image {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 913.304797] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Creating directory with path [datastore1] devstack-image-cache_base/02de5fc5-60f6-49e2-ae03-e14c840dae1b {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 913.305521] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3b8f5aa4-50e8-413a-8e87-609d4a15285b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.312489] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.493s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.312489] env[63021]: DEBUG nova.compute.manager [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 913.314230] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 8.933s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.330120] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Created directory with path [datastore1] devstack-image-cache_base/02de5fc5-60f6-49e2-ae03-e14c840dae1b {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 913.330120] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_00c60d10-c465-4c17-b188-9861d55eeda2/OSTACK_IMG_00c60d10-c465-4c17-b188-9861d55eeda2.vmdk to [datastore1] devstack-image-cache_base/02de5fc5-60f6-49e2-ae03-e14c840dae1b/02de5fc5-60f6-49e2-ae03-e14c840dae1b.vmdk. {{(pid=63021) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 913.330636] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-80bc7c7c-9ddb-4357-ac4a-7f44187264b0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.339529] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 913.339529] env[63021]: value = "task-1293943" [ 913.339529] env[63021]: _type = "Task" [ 913.339529] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.349861] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293943, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.565355] env[63021]: DEBUG oslo_concurrency.lockutils [None req-96ce6cf7-19a2-4299-affa-85e2a3ac4827 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Releasing lock "refresh_cache-df856146-61ba-4de7-a5fd-d75263927163" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.580611] env[63021]: DEBUG oslo_vmware.api [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293940, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.587960] env[63021]: DEBUG nova.compute.manager [req-746c078d-1d15-4f1f-9828-fac734bb43ab req-6969bc7b-c7d3-40af-bde6-d86ee1e7a80f service nova] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Received event network-vif-deleted-71541497-0e60-4ea5-9d4a-33970a716736 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 913.588249] env[63021]: DEBUG nova.compute.manager [req-746c078d-1d15-4f1f-9828-fac734bb43ab req-6969bc7b-c7d3-40af-bde6-d86ee1e7a80f service nova] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Received event network-vif-deleted-9ba9d35e-148e-45b3-a105-3ed457c3803c {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 913.618503] env[63021]: DEBUG oslo_vmware.api [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293941, 'name': PowerOffVM_Task, 'duration_secs': 0.20952} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.619436] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 913.619675] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 913.619966] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ce2de720-164a-4bf4-a4ce-e85ff48c6ad0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.638190] env[63021]: DEBUG nova.objects.instance [None req-02bb7a1f-0d68-4e1f-8b5b-41cd7beaf066 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lazy-loading 'flavor' on Instance uuid c328d2f7-3398-4f25-b11c-f464be7af8a1 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 913.689692] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 913.689940] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Deleting contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 913.691058] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Deleting the datastore file [datastore2] 1b4871cc-3dac-4683-8fce-93c4d9e4407d {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 913.691058] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-79f8219e-6841-426e-bb8b-1c96725ff20c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.701197] env[63021]: DEBUG oslo_vmware.api [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 913.701197] env[63021]: value = "task-1293945" [ 913.701197] env[63021]: _type = "Task" [ 913.701197] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.711220] env[63021]: DEBUG oslo_vmware.api [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293945, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.818512] env[63021]: DEBUG nova.compute.utils [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 913.835300] env[63021]: DEBUG nova.compute.manager [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 913.839849] env[63021]: DEBUG nova.compute.manager [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 913.840492] env[63021]: DEBUG nova.network.neutron [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 913.866263] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293943, 'name': MoveVirtualDisk_Task} progress is 12%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.922169] env[63021]: DEBUG nova.policy [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b427f7d1317341f38754a2b324f4f5aa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e238d5c1668540d2a9fffd3fd832d9b8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 913.996502] env[63021]: DEBUG nova.network.neutron [req-c568f9b6-4edb-430a-bc9a-e8ceede4d0f7 req-0f3fe796-c71c-44a8-a831-0a893a133e16 service nova] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Updated VIF entry in instance network info cache for port 2e8a7fc6-825c-45a5-b41c-d28410061aa3. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 913.997200] env[63021]: DEBUG nova.network.neutron [req-c568f9b6-4edb-430a-bc9a-e8ceede4d0f7 req-0f3fe796-c71c-44a8-a831-0a893a133e16 service nova] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Updating instance_info_cache with network_info: [{"id": "2e8a7fc6-825c-45a5-b41c-d28410061aa3", "address": "fa:16:3e:61:15:99", "network": {"id": "f18b6fd7-dc66-48f7-a183-e22056ae2daf", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-748006433-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.190", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9a23f10cf5a4b178c7efff86041ce21", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e8a7fc6-82", "ovs_interfaceid": "2e8a7fc6-825c-45a5-b41c-d28410061aa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.072268] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-96ce6cf7-19a2-4299-affa-85e2a3ac4827 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 914.072923] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-68a1dec2-15d8-4504-afc8-9051b0451b36 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.089021] env[63021]: DEBUG oslo_vmware.api [None req-96ce6cf7-19a2-4299-affa-85e2a3ac4827 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 914.089021] env[63021]: value = "task-1293946" [ 914.089021] env[63021]: _type = "Task" [ 914.089021] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.089497] env[63021]: DEBUG oslo_vmware.api [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293940, 'name': ReconfigVM_Task, 'duration_secs': 0.658131} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.092907] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Reconfigured VM instance instance-0000004e to attach disk [datastore1] 4d712cc1-6202-407e-8767-c620a1ce0bba/4d712cc1-6202-407e-8767-c620a1ce0bba.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 914.094446] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-add61d54-3c75-478f-afdf-d54d7512c467 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.106740] env[63021]: DEBUG oslo_vmware.api [None req-96ce6cf7-19a2-4299-affa-85e2a3ac4827 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1293946, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.106740] env[63021]: DEBUG oslo_vmware.api [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Waiting for the task: (returnval){ [ 914.106740] env[63021]: value = "task-1293947" [ 914.106740] env[63021]: _type = "Task" [ 914.106740] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.116379] env[63021]: DEBUG oslo_vmware.api [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293947, 'name': Rename_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.144239] env[63021]: DEBUG oslo_concurrency.lockutils [None req-02bb7a1f-0d68-4e1f-8b5b-41cd7beaf066 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "refresh_cache-c328d2f7-3398-4f25-b11c-f464be7af8a1" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.144547] env[63021]: DEBUG oslo_concurrency.lockutils [None req-02bb7a1f-0d68-4e1f-8b5b-41cd7beaf066 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquired lock "refresh_cache-c328d2f7-3398-4f25-b11c-f464be7af8a1" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.144860] env[63021]: DEBUG nova.network.neutron [None req-02bb7a1f-0d68-4e1f-8b5b-41cd7beaf066 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 914.145055] env[63021]: DEBUG nova.objects.instance [None req-02bb7a1f-0d68-4e1f-8b5b-41cd7beaf066 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lazy-loading 'info_cache' on Instance uuid c328d2f7-3398-4f25-b11c-f464be7af8a1 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 914.211542] env[63021]: DEBUG oslo_vmware.api [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293945, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.419608} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.212283] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 914.212283] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Deleted contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 914.212621] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 914.212919] env[63021]: INFO nova.compute.manager [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Took 1.12 seconds to destroy the instance on the hypervisor. [ 914.213347] env[63021]: DEBUG oslo.service.loopingcall [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 914.213724] env[63021]: DEBUG nova.compute.manager [-] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 914.215674] env[63021]: DEBUG nova.network.neutron [-] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 914.300033] env[63021]: DEBUG nova.network.neutron [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Successfully created port: 89471183-ed50-4a4d-85ce-790ebdf37c27 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 914.351270] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Applying migration context for instance 71909213-258d-45f5-9a3e-e473e8ce8aa2 as it has an incoming, in-progress migration b36d8c79-3341-4688-9479-6846fc53dcad. Migration status is confirming {{(pid=63021) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 914.353194] env[63021]: INFO nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Updating resource usage from migration b36d8c79-3341-4688-9479-6846fc53dcad [ 914.372027] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293943, 'name': MoveVirtualDisk_Task} progress is 32%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.378772] env[63021]: WARNING nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 7c6d7839-616d-41f5-a909-f8c626477a5e is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 914.379094] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance df0b6974-7f78-4b73-8583-d09754fbacea actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 914.379329] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance df856146-61ba-4de7-a5fd-d75263927163 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 914.379552] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 45c9e2be-eb19-4b83-b280-c9eeaddfccba actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 914.379768] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance c328d2f7-3398-4f25-b11c-f464be7af8a1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 914.379973] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 881c2fb8-5f8c-48c2-a173-8fab949f24f7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 914.380212] env[63021]: WARNING nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 44dadf8e-a727-448f-887d-6408f7ebfbc6 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 914.380966] env[63021]: WARNING nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 914.380966] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 1b4871cc-3dac-4683-8fce-93c4d9e4407d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 914.380966] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 80b857c4-d9dd-4483-970f-a0f5f4bcf173 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 914.380966] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 4d712cc1-6202-407e-8767-c620a1ce0bba actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 914.380966] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Migration b36d8c79-3341-4688-9479-6846fc53dcad is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 914.381360] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 71909213-258d-45f5-9a3e-e473e8ce8aa2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 914.381649] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance ed26dff1-fed0-4baf-ad41-d14850254aec actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 914.383160] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance d63fd465-975f-42ce-b1be-ac8a6929bc4d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 914.383160] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 140184cd-6e4c-425c-8b17-361a1f565e93 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 914.501655] env[63021]: DEBUG oslo_concurrency.lockutils [req-c568f9b6-4edb-430a-bc9a-e8ceede4d0f7 req-0f3fe796-c71c-44a8-a831-0a893a133e16 service nova] Releasing lock "refresh_cache-ed26dff1-fed0-4baf-ad41-d14850254aec" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.604276] env[63021]: DEBUG oslo_vmware.api [None req-96ce6cf7-19a2-4299-affa-85e2a3ac4827 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1293946, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.617440] env[63021]: DEBUG oslo_vmware.api [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293947, 'name': Rename_Task, 'duration_secs': 0.45095} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.617936] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 914.618290] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-df15fa68-1706-4c7b-adb2-ffe3f96ddf7d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.625150] env[63021]: DEBUG oslo_vmware.api [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Waiting for the task: (returnval){ [ 914.625150] env[63021]: value = "task-1293948" [ 914.625150] env[63021]: _type = "Task" [ 914.625150] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.632624] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "2d647028-306a-41d5-96f0-2e82a98ab56d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.632868] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "2d647028-306a-41d5-96f0-2e82a98ab56d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.642841] env[63021]: DEBUG oslo_vmware.api [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293948, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.649455] env[63021]: DEBUG nova.objects.base [None req-02bb7a1f-0d68-4e1f-8b5b-41cd7beaf066 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=63021) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 914.870969] env[63021]: DEBUG nova.compute.manager [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 914.873933] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293943, 'name': MoveVirtualDisk_Task} progress is 49%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.884733] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance ae103118-bddf-46fe-90b9-98b60952ebba has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 914.901164] env[63021]: DEBUG nova.virt.hardware [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 914.901682] env[63021]: DEBUG nova.virt.hardware [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 914.901883] env[63021]: DEBUG nova.virt.hardware [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 914.902357] env[63021]: DEBUG nova.virt.hardware [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 914.902616] env[63021]: DEBUG nova.virt.hardware [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 914.902847] env[63021]: DEBUG nova.virt.hardware [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 914.903193] env[63021]: DEBUG nova.virt.hardware [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 914.903554] env[63021]: DEBUG nova.virt.hardware [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 914.903812] env[63021]: DEBUG nova.virt.hardware [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 914.904084] env[63021]: DEBUG nova.virt.hardware [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 914.904455] env[63021]: DEBUG nova.virt.hardware [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 914.906698] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eb29e8f-c262-4045-a795-342ed31f7c34 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.917280] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35d0b7bc-9a50-4cca-816a-efbfa64f010a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.100034] env[63021]: DEBUG oslo_vmware.api [None req-96ce6cf7-19a2-4299-affa-85e2a3ac4827 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1293946, 'name': PowerOnVM_Task, 'duration_secs': 0.544567} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.100358] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-96ce6cf7-19a2-4299-affa-85e2a3ac4827 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 915.100584] env[63021]: DEBUG nova.compute.manager [None req-96ce6cf7-19a2-4299-affa-85e2a3ac4827 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 915.101458] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7fbcd76-6ba9-433c-9bec-93484e550f84 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.136156] env[63021]: DEBUG nova.compute.manager [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 915.139461] env[63021]: DEBUG oslo_vmware.api [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293948, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.358334] env[63021]: DEBUG nova.network.neutron [-] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.375679] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293943, 'name': MoveVirtualDisk_Task} progress is 69%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.388512] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 418a394b-dbb2-47a2-96a9-a1caa4c4bb51 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 915.421623] env[63021]: DEBUG nova.network.neutron [None req-02bb7a1f-0d68-4e1f-8b5b-41cd7beaf066 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Updating instance_info_cache with network_info: [{"id": "eff6f35a-0a13-4e0c-88b0-96b66b240f9b", "address": "fa:16:3e:27:3b:5e", "network": {"id": "9502e364-60e1-47ef-a1f7-f3e3ff45cf69", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-164717784-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5af87713ae6b477ca8a584aed287889d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeff6f35a-0a", "ovs_interfaceid": "eff6f35a-0a13-4e0c-88b0-96b66b240f9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.615591] env[63021]: DEBUG nova.compute.manager [req-e18ae346-68fd-4582-81f2-1ca9313c5222 req-310f8bb6-3af5-4564-a1b0-475b43a6cb21 service nova] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Received event network-vif-deleted-d34cfd0d-464e-400b-8366-83321cc6112f {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 915.637621] env[63021]: DEBUG oslo_vmware.api [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293948, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.659158] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.861491] env[63021]: INFO nova.compute.manager [-] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Took 1.65 seconds to deallocate network for instance. [ 915.878035] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293943, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.892092] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 2d647028-306a-41d5-96f0-2e82a98ab56d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 915.892486] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Total usable vcpus: 48, total allocated vcpus: 13 {{(pid=63021) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 915.892608] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3072MB phys_disk=200GB used_disk=12GB total_vcpus=48 used_vcpus=13 pci_stats=[] {{(pid=63021) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 915.924608] env[63021]: DEBUG oslo_concurrency.lockutils [None req-02bb7a1f-0d68-4e1f-8b5b-41cd7beaf066 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Releasing lock "refresh_cache-c328d2f7-3398-4f25-b11c-f464be7af8a1" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.140787] env[63021]: DEBUG oslo_vmware.api [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293948, 'name': PowerOnVM_Task, 'duration_secs': 1.275172} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.141171] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 916.141391] env[63021]: DEBUG nova.compute.manager [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 916.142289] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-382b06c7-f37d-4931-99c1-4b715e47d5e4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.173174] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53923199-a6ed-461e-b61c-47e68cf088a3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.181672] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4639cb4-29a0-4a9a-bb5c-0d00f8b6f79e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.217085] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-739d3638-a014-4eb2-8a4c-982ae4e4385d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.225545] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34a69b7c-c6dd-40ca-a449-a267bd07ee37 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.229440] env[63021]: DEBUG nova.network.neutron [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Successfully updated port: 89471183-ed50-4a4d-85ce-790ebdf37c27 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 916.240860] env[63021]: DEBUG nova.compute.provider_tree [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 916.374971] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.375096] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293943, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.427873] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-02bb7a1f-0d68-4e1f-8b5b-41cd7beaf066 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 916.428436] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cb8f9f48-4810-484e-97fa-ade1f0b2df29 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.436145] env[63021]: DEBUG oslo_vmware.api [None req-02bb7a1f-0d68-4e1f-8b5b-41cd7beaf066 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 916.436145] env[63021]: value = "task-1293949" [ 916.436145] env[63021]: _type = "Task" [ 916.436145] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.444751] env[63021]: DEBUG oslo_vmware.api [None req-02bb7a1f-0d68-4e1f-8b5b-41cd7beaf066 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1293949, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.661317] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.742872] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "refresh_cache-140184cd-6e4c-425c-8b17-361a1f565e93" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.743094] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquired lock "refresh_cache-140184cd-6e4c-425c-8b17-361a1f565e93" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.743345] env[63021]: DEBUG nova.network.neutron [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 916.745073] env[63021]: DEBUG nova.scheduler.client.report [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 916.875142] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293943, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.231509} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.875560] env[63021]: INFO nova.virt.vmwareapi.ds_util [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_00c60d10-c465-4c17-b188-9861d55eeda2/OSTACK_IMG_00c60d10-c465-4c17-b188-9861d55eeda2.vmdk to [datastore1] devstack-image-cache_base/02de5fc5-60f6-49e2-ae03-e14c840dae1b/02de5fc5-60f6-49e2-ae03-e14c840dae1b.vmdk. [ 916.875787] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Cleaning up location [datastore1] OSTACK_IMG_00c60d10-c465-4c17-b188-9861d55eeda2 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 916.875978] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_00c60d10-c465-4c17-b188-9861d55eeda2 {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 916.876293] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-90e61154-0feb-49ef-89df-4c25bf6630c4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.882733] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 916.882733] env[63021]: value = "task-1293950" [ 916.882733] env[63021]: _type = "Task" [ 916.882733] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.890162] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293950, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.949458] env[63021]: DEBUG oslo_vmware.api [None req-02bb7a1f-0d68-4e1f-8b5b-41cd7beaf066 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1293949, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.195979] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Acquiring lock "4d712cc1-6202-407e-8767-c620a1ce0bba" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.195979] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Lock "4d712cc1-6202-407e-8767-c620a1ce0bba" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.195979] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Acquiring lock "4d712cc1-6202-407e-8767-c620a1ce0bba-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.195979] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Lock "4d712cc1-6202-407e-8767-c620a1ce0bba-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.195979] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Lock "4d712cc1-6202-407e-8767-c620a1ce0bba-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.198197] env[63021]: INFO nova.compute.manager [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Terminating instance [ 917.199981] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Acquiring lock "refresh_cache-4d712cc1-6202-407e-8767-c620a1ce0bba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.200158] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Acquired lock "refresh_cache-4d712cc1-6202-407e-8767-c620a1ce0bba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.200350] env[63021]: DEBUG nova.network.neutron [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 917.251025] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63021) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 917.251311] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.937s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.251601] env[63021]: DEBUG oslo_concurrency.lockutils [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.406s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.253204] env[63021]: INFO nova.compute.claims [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 917.288275] env[63021]: DEBUG nova.network.neutron [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 917.397150] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293950, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.116394} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.399885] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 917.400108] env[63021]: DEBUG oslo_concurrency.lockutils [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02de5fc5-60f6-49e2-ae03-e14c840dae1b/02de5fc5-60f6-49e2-ae03-e14c840dae1b.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.400382] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02de5fc5-60f6-49e2-ae03-e14c840dae1b/02de5fc5-60f6-49e2-ae03-e14c840dae1b.vmdk to [datastore1] d63fd465-975f-42ce-b1be-ac8a6929bc4d/d63fd465-975f-42ce-b1be-ac8a6929bc4d.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 917.400679] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-56f8b919-66fa-4cef-bb30-d7c5cb42232d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.408952] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 917.408952] env[63021]: value = "task-1293951" [ 917.408952] env[63021]: _type = "Task" [ 917.408952] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.416298] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293951, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.437649] env[63021]: DEBUG nova.network.neutron [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Updating instance_info_cache with network_info: [{"id": "89471183-ed50-4a4d-85ce-790ebdf37c27", "address": "fa:16:3e:bc:46:c6", "network": {"id": "81917a55-4e41-4898-9f4e-4fced52b5a86", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1793837736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e238d5c1668540d2a9fffd3fd832d9b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89471183-ed", "ovs_interfaceid": "89471183-ed50-4a4d-85ce-790ebdf37c27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.447553] env[63021]: DEBUG oslo_vmware.api [None req-02bb7a1f-0d68-4e1f-8b5b-41cd7beaf066 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1293949, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.646490] env[63021]: DEBUG nova.compute.manager [req-f4567eb5-93ce-41bb-a939-238f19f510b2 req-da0cc537-3660-4aa2-b20e-306792ec1e71 service nova] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Received event network-vif-plugged-89471183-ed50-4a4d-85ce-790ebdf37c27 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 917.646850] env[63021]: DEBUG oslo_concurrency.lockutils [req-f4567eb5-93ce-41bb-a939-238f19f510b2 req-da0cc537-3660-4aa2-b20e-306792ec1e71 service nova] Acquiring lock "140184cd-6e4c-425c-8b17-361a1f565e93-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.647627] env[63021]: DEBUG oslo_concurrency.lockutils [req-f4567eb5-93ce-41bb-a939-238f19f510b2 req-da0cc537-3660-4aa2-b20e-306792ec1e71 service nova] Lock "140184cd-6e4c-425c-8b17-361a1f565e93-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.648071] env[63021]: DEBUG oslo_concurrency.lockutils [req-f4567eb5-93ce-41bb-a939-238f19f510b2 req-da0cc537-3660-4aa2-b20e-306792ec1e71 service nova] Lock "140184cd-6e4c-425c-8b17-361a1f565e93-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.648246] env[63021]: DEBUG nova.compute.manager [req-f4567eb5-93ce-41bb-a939-238f19f510b2 req-da0cc537-3660-4aa2-b20e-306792ec1e71 service nova] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] No waiting events found dispatching network-vif-plugged-89471183-ed50-4a4d-85ce-790ebdf37c27 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 917.648541] env[63021]: WARNING nova.compute.manager [req-f4567eb5-93ce-41bb-a939-238f19f510b2 req-da0cc537-3660-4aa2-b20e-306792ec1e71 service nova] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Received unexpected event network-vif-plugged-89471183-ed50-4a4d-85ce-790ebdf37c27 for instance with vm_state building and task_state spawning. [ 917.648834] env[63021]: DEBUG nova.compute.manager [req-f4567eb5-93ce-41bb-a939-238f19f510b2 req-da0cc537-3660-4aa2-b20e-306792ec1e71 service nova] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Received event network-changed-89471183-ed50-4a4d-85ce-790ebdf37c27 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 917.652959] env[63021]: DEBUG nova.compute.manager [req-f4567eb5-93ce-41bb-a939-238f19f510b2 req-da0cc537-3660-4aa2-b20e-306792ec1e71 service nova] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Refreshing instance network info cache due to event network-changed-89471183-ed50-4a4d-85ce-790ebdf37c27. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 917.653324] env[63021]: DEBUG oslo_concurrency.lockutils [req-f4567eb5-93ce-41bb-a939-238f19f510b2 req-da0cc537-3660-4aa2-b20e-306792ec1e71 service nova] Acquiring lock "refresh_cache-140184cd-6e4c-425c-8b17-361a1f565e93" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.719201] env[63021]: DEBUG nova.network.neutron [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 917.771656] env[63021]: DEBUG nova.network.neutron [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.919576] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293951, 'name': CopyVirtualDisk_Task} progress is 15%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.944966] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Releasing lock "refresh_cache-140184cd-6e4c-425c-8b17-361a1f565e93" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.945325] env[63021]: DEBUG nova.compute.manager [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Instance network_info: |[{"id": "89471183-ed50-4a4d-85ce-790ebdf37c27", "address": "fa:16:3e:bc:46:c6", "network": {"id": "81917a55-4e41-4898-9f4e-4fced52b5a86", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1793837736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e238d5c1668540d2a9fffd3fd832d9b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89471183-ed", "ovs_interfaceid": "89471183-ed50-4a4d-85ce-790ebdf37c27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 917.945648] env[63021]: DEBUG oslo_concurrency.lockutils [req-f4567eb5-93ce-41bb-a939-238f19f510b2 req-da0cc537-3660-4aa2-b20e-306792ec1e71 service nova] Acquired lock "refresh_cache-140184cd-6e4c-425c-8b17-361a1f565e93" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.945922] env[63021]: DEBUG nova.network.neutron [req-f4567eb5-93ce-41bb-a939-238f19f510b2 req-da0cc537-3660-4aa2-b20e-306792ec1e71 service nova] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Refreshing network info cache for port 89471183-ed50-4a4d-85ce-790ebdf37c27 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 917.947137] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bc:46:c6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '535b175f-71d3-4226-81ab-ca253f27fedd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '89471183-ed50-4a4d-85ce-790ebdf37c27', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 917.954530] env[63021]: DEBUG oslo.service.loopingcall [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 917.957564] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 917.960875] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3c32c816-74c3-4fb5-9955-3fe2757ed1c5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.976038] env[63021]: DEBUG oslo_vmware.api [None req-02bb7a1f-0d68-4e1f-8b5b-41cd7beaf066 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1293949, 'name': PowerOnVM_Task, 'duration_secs': 1.015277} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.976319] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-02bb7a1f-0d68-4e1f-8b5b-41cd7beaf066 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 917.976537] env[63021]: DEBUG nova.compute.manager [None req-02bb7a1f-0d68-4e1f-8b5b-41cd7beaf066 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 917.977750] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2de271ef-be8c-4087-953d-92c19d5f6456 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.982519] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 917.982519] env[63021]: value = "task-1293952" [ 917.982519] env[63021]: _type = "Task" [ 917.982519] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.993691] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293952, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.187104] env[63021]: DEBUG nova.network.neutron [req-f4567eb5-93ce-41bb-a939-238f19f510b2 req-da0cc537-3660-4aa2-b20e-306792ec1e71 service nova] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Updated VIF entry in instance network info cache for port 89471183-ed50-4a4d-85ce-790ebdf37c27. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 918.187580] env[63021]: DEBUG nova.network.neutron [req-f4567eb5-93ce-41bb-a939-238f19f510b2 req-da0cc537-3660-4aa2-b20e-306792ec1e71 service nova] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Updating instance_info_cache with network_info: [{"id": "89471183-ed50-4a4d-85ce-790ebdf37c27", "address": "fa:16:3e:bc:46:c6", "network": {"id": "81917a55-4e41-4898-9f4e-4fced52b5a86", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1793837736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e238d5c1668540d2a9fffd3fd832d9b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89471183-ed", "ovs_interfaceid": "89471183-ed50-4a4d-85ce-790ebdf37c27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.274970] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Releasing lock "refresh_cache-4d712cc1-6202-407e-8767-c620a1ce0bba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.275506] env[63021]: DEBUG nova.compute.manager [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 918.275757] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 918.276813] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3473b0f7-b427-498b-b459-ebc506080fb6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.285024] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 918.287761] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a3c1c631-8a8e-496b-8ed5-2cbcb6b64d86 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.293381] env[63021]: DEBUG oslo_vmware.api [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Waiting for the task: (returnval){ [ 918.293381] env[63021]: value = "task-1293953" [ 918.293381] env[63021]: _type = "Task" [ 918.293381] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.303031] env[63021]: DEBUG oslo_vmware.api [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293953, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.421107] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293951, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.495452] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293952, 'name': CreateVM_Task, 'duration_secs': 0.357773} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.496795] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 918.497388] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.497590] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.497907] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 918.498361] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f00cb58a-2575-4c83-85cd-c5ddf57c1d50 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.506205] env[63021]: DEBUG oslo_vmware.api [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 918.506205] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]529f6599-a14f-d348-c8e1-23b67ddeefc8" [ 918.506205] env[63021]: _type = "Task" [ 918.506205] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.517463] env[63021]: DEBUG oslo_vmware.api [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]529f6599-a14f-d348-c8e1-23b67ddeefc8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.528040] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3752e46-8457-4937-ad43-fab18314fefb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.535073] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c630ce99-e858-4d51-9a12-a8d97d3f277b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.566052] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54818809-1b6a-43da-b10f-e0644f35129c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.573137] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e065b1c6-a1ca-452a-8027-e0e64fbb79bb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.586334] env[63021]: DEBUG nova.compute.provider_tree [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 918.690535] env[63021]: DEBUG oslo_concurrency.lockutils [req-f4567eb5-93ce-41bb-a939-238f19f510b2 req-da0cc537-3660-4aa2-b20e-306792ec1e71 service nova] Releasing lock "refresh_cache-140184cd-6e4c-425c-8b17-361a1f565e93" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.803751] env[63021]: DEBUG oslo_vmware.api [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293953, 'name': PowerOffVM_Task, 'duration_secs': 0.227465} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.804112] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 918.804323] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 918.804624] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c1edf8ec-d752-4873-a31a-25fe78ddddc1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.829264] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 918.829482] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 918.829671] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Deleting the datastore file [datastore1] 4d712cc1-6202-407e-8767-c620a1ce0bba {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 918.829942] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bff4dc35-f4d9-4504-aaf4-f77c7dd82f4a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.835562] env[63021]: DEBUG oslo_vmware.api [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Waiting for the task: (returnval){ [ 918.835562] env[63021]: value = "task-1293955" [ 918.835562] env[63021]: _type = "Task" [ 918.835562] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.844791] env[63021]: DEBUG oslo_vmware.api [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293955, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.922538] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293951, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.016114] env[63021]: DEBUG oslo_vmware.api [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]529f6599-a14f-d348-c8e1-23b67ddeefc8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.090043] env[63021]: DEBUG nova.scheduler.client.report [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 919.345611] env[63021]: DEBUG oslo_vmware.api [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293955, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.421135] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293951, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.516785] env[63021]: DEBUG oslo_vmware.api [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]529f6599-a14f-d348-c8e1-23b67ddeefc8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.595125] env[63021]: DEBUG oslo_concurrency.lockutils [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.343s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.595793] env[63021]: DEBUG nova.compute.manager [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 919.598379] env[63021]: DEBUG oslo_concurrency.lockutils [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.874s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.600069] env[63021]: INFO nova.compute.claims [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 919.847962] env[63021]: DEBUG oslo_vmware.api [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293955, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.922060] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293951, 'name': CopyVirtualDisk_Task} progress is 29%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.018288] env[63021]: DEBUG oslo_vmware.api [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]529f6599-a14f-d348-c8e1-23b67ddeefc8, 'name': SearchDatastore_Task, 'duration_secs': 1.263331} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.018627] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.018923] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 920.019195] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.019349] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.019603] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 920.019909] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-68b872fc-d4ab-4fe1-9f68-9167a78c14d2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.036499] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 920.036732] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 920.037597] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-23de7ca4-f896-49e8-abb9-ccf18188b4c5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.044033] env[63021]: DEBUG oslo_vmware.api [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 920.044033] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5216efe1-dc88-96fa-e035-0a3594cda373" [ 920.044033] env[63021]: _type = "Task" [ 920.044033] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.052490] env[63021]: DEBUG oslo_vmware.api [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5216efe1-dc88-96fa-e035-0a3594cda373, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.104461] env[63021]: DEBUG nova.compute.utils [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 920.108521] env[63021]: DEBUG nova.compute.manager [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 920.108668] env[63021]: DEBUG nova.network.neutron [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 920.150507] env[63021]: DEBUG nova.policy [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6f9dc7ceda4f4ddbba1de296ee4cc814', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b672b1d7d7ad454ebfa3e2fd2421c754', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 920.348249] env[63021]: DEBUG oslo_vmware.api [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293955, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.425701] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293951, 'name': CopyVirtualDisk_Task} progress is 52%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.443810] env[63021]: DEBUG nova.network.neutron [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Successfully created port: f99a201d-df02-499f-85d9-2566ac1dd82b {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 920.558011] env[63021]: DEBUG oslo_vmware.api [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5216efe1-dc88-96fa-e035-0a3594cda373, 'name': SearchDatastore_Task, 'duration_secs': 0.085625} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.558921] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ef9babd-37d1-45bf-920a-f3bb5ef8dcc9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.564715] env[63021]: DEBUG oslo_vmware.api [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 920.564715] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5271a12a-8b48-8ad0-905e-39db17067f95" [ 920.564715] env[63021]: _type = "Task" [ 920.564715] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.574500] env[63021]: DEBUG oslo_vmware.api [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5271a12a-8b48-8ad0-905e-39db17067f95, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.609640] env[63021]: DEBUG nova.compute.manager [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 920.847204] env[63021]: DEBUG oslo_vmware.api [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293955, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.887446] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad485e93-3181-43b0-9aee-28d423832718 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.895265] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b9b4dcd-767c-48d4-9fd8-cb7c8f50c44e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.929432] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-357b1c26-5094-4057-b5e8-0f2d1ae9df1e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.937549] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293951, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.940844] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e52e04a-cda7-40c0-8a90-6079b8badd7b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.958335] env[63021]: DEBUG nova.compute.provider_tree [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 921.078282] env[63021]: DEBUG oslo_vmware.api [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5271a12a-8b48-8ad0-905e-39db17067f95, 'name': SearchDatastore_Task, 'duration_secs': 0.094212} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.079160] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.079160] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 140184cd-6e4c-425c-8b17-361a1f565e93/140184cd-6e4c-425c-8b17-361a1f565e93.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 921.079284] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a8ce9a7d-aed2-4ed5-8b6a-7013bc0783c5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.089117] env[63021]: DEBUG oslo_vmware.api [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 921.089117] env[63021]: value = "task-1293956" [ 921.089117] env[63021]: _type = "Task" [ 921.089117] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.101899] env[63021]: DEBUG oslo_vmware.api [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293956, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.350698] env[63021]: DEBUG oslo_vmware.api [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Task: {'id': task-1293955, 'name': DeleteDatastoreFile_Task, 'duration_secs': 2.076304} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.351094] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 921.351369] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 921.351548] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 921.351745] env[63021]: INFO nova.compute.manager [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Took 3.08 seconds to destroy the instance on the hypervisor. [ 921.352009] env[63021]: DEBUG oslo.service.loopingcall [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 921.352253] env[63021]: DEBUG nova.compute.manager [-] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 921.352371] env[63021]: DEBUG nova.network.neutron [-] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 921.370424] env[63021]: DEBUG nova.network.neutron [-] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 921.437428] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293951, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.462115] env[63021]: DEBUG nova.scheduler.client.report [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 921.602532] env[63021]: DEBUG oslo_vmware.api [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293956, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.621747] env[63021]: DEBUG nova.compute.manager [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 921.650610] env[63021]: DEBUG nova.virt.hardware [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 921.650904] env[63021]: DEBUG nova.virt.hardware [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 921.651175] env[63021]: DEBUG nova.virt.hardware [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 921.651435] env[63021]: DEBUG nova.virt.hardware [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 921.651672] env[63021]: DEBUG nova.virt.hardware [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 921.651914] env[63021]: DEBUG nova.virt.hardware [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 921.652227] env[63021]: DEBUG nova.virt.hardware [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 921.652414] env[63021]: DEBUG nova.virt.hardware [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 921.652692] env[63021]: DEBUG nova.virt.hardware [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 921.652898] env[63021]: DEBUG nova.virt.hardware [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 921.653149] env[63021]: DEBUG nova.virt.hardware [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 921.654193] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13b56c86-85a2-40cb-a718-809d0b3d9654 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.666524] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20b542da-d7ed-42af-ab58-357ee709d360 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.874025] env[63021]: DEBUG nova.network.neutron [-] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.880690] env[63021]: DEBUG nova.compute.manager [req-c24c3d2a-b5f4-4823-817b-f833ea449402 req-20875df8-4512-4378-8f55-c22267eac9e3 service nova] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Received event network-vif-plugged-f99a201d-df02-499f-85d9-2566ac1dd82b {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 921.880982] env[63021]: DEBUG oslo_concurrency.lockutils [req-c24c3d2a-b5f4-4823-817b-f833ea449402 req-20875df8-4512-4378-8f55-c22267eac9e3 service nova] Acquiring lock "ae103118-bddf-46fe-90b9-98b60952ebba-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.881602] env[63021]: DEBUG oslo_concurrency.lockutils [req-c24c3d2a-b5f4-4823-817b-f833ea449402 req-20875df8-4512-4378-8f55-c22267eac9e3 service nova] Lock "ae103118-bddf-46fe-90b9-98b60952ebba-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.881602] env[63021]: DEBUG oslo_concurrency.lockutils [req-c24c3d2a-b5f4-4823-817b-f833ea449402 req-20875df8-4512-4378-8f55-c22267eac9e3 service nova] Lock "ae103118-bddf-46fe-90b9-98b60952ebba-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.881733] env[63021]: DEBUG nova.compute.manager [req-c24c3d2a-b5f4-4823-817b-f833ea449402 req-20875df8-4512-4378-8f55-c22267eac9e3 service nova] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] No waiting events found dispatching network-vif-plugged-f99a201d-df02-499f-85d9-2566ac1dd82b {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 921.881993] env[63021]: WARNING nova.compute.manager [req-c24c3d2a-b5f4-4823-817b-f833ea449402 req-20875df8-4512-4378-8f55-c22267eac9e3 service nova] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Received unexpected event network-vif-plugged-f99a201d-df02-499f-85d9-2566ac1dd82b for instance with vm_state building and task_state spawning. [ 921.938366] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293951, 'name': CopyVirtualDisk_Task, 'duration_secs': 4.317578} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.938631] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02de5fc5-60f6-49e2-ae03-e14c840dae1b/02de5fc5-60f6-49e2-ae03-e14c840dae1b.vmdk to [datastore1] d63fd465-975f-42ce-b1be-ac8a6929bc4d/d63fd465-975f-42ce-b1be-ac8a6929bc4d.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 921.939488] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2abb9c01-b3f4-42ed-869b-126d6437a9c9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.970758] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] d63fd465-975f-42ce-b1be-ac8a6929bc4d/d63fd465-975f-42ce-b1be-ac8a6929bc4d.vmdk or device None with type streamOptimized {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 921.970758] env[63021]: DEBUG oslo_concurrency.lockutils [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.370s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.970758] env[63021]: DEBUG nova.compute.manager [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 921.972194] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dcf4eb98-aff2-4d97-bb4f-9a01fd324a73 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.988517] env[63021]: DEBUG nova.network.neutron [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Successfully updated port: f99a201d-df02-499f-85d9-2566ac1dd82b {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 921.989944] env[63021]: DEBUG oslo_concurrency.lockutils [None req-65a46cd5-0e5c-4e9c-bb88-b7b7e09d89a7 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 11.058s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.999852] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 921.999852] env[63021]: value = "task-1293957" [ 921.999852] env[63021]: _type = "Task" [ 921.999852] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.014234] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293957, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.103877] env[63021]: DEBUG oslo_vmware.api [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293956, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.375920] env[63021]: INFO nova.compute.manager [-] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Took 1.02 seconds to deallocate network for instance. [ 922.490507] env[63021]: DEBUG nova.compute.utils [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 922.492336] env[63021]: DEBUG nova.compute.manager [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 922.492483] env[63021]: DEBUG nova.network.neutron [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 922.494877] env[63021]: DEBUG oslo_concurrency.lockutils [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "refresh_cache-ae103118-bddf-46fe-90b9-98b60952ebba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.495035] env[63021]: DEBUG oslo_concurrency.lockutils [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquired lock "refresh_cache-ae103118-bddf-46fe-90b9-98b60952ebba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.495180] env[63021]: DEBUG nova.network.neutron [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 922.512888] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293957, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.543789] env[63021]: DEBUG nova.policy [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6f9dc7ceda4f4ddbba1de296ee4cc814', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b672b1d7d7ad454ebfa3e2fd2421c754', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 922.603213] env[63021]: DEBUG oslo_vmware.api [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293956, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.166342} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.603507] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 140184cd-6e4c-425c-8b17-361a1f565e93/140184cd-6e4c-425c-8b17-361a1f565e93.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 922.603907] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 922.604069] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fbb6c1db-a1b2-4ef6-9676-d48357117bbc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.610539] env[63021]: DEBUG oslo_vmware.api [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 922.610539] env[63021]: value = "task-1293958" [ 922.610539] env[63021]: _type = "Task" [ 922.610539] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.620823] env[63021]: DEBUG oslo_vmware.api [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293958, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.800971] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de671403-b238-4cbd-98c4-857c3613f753 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.809300] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ca17720-44c0-4891-b91a-94f365e48a25 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.844398] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c9b4181-8d51-4bf0-a58b-39576647a4f6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.850925] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e85c107-005a-49e3-8dc9-655c95688568 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.868048] env[63021]: DEBUG nova.compute.provider_tree [None req-65a46cd5-0e5c-4e9c-bb88-b7b7e09d89a7 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 922.875631] env[63021]: DEBUG nova.network.neutron [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Successfully created port: 905466dc-1047-469f-b282-768293b6d154 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 922.883341] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.000740] env[63021]: DEBUG nova.compute.manager [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 923.017194] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293957, 'name': ReconfigVM_Task, 'duration_secs': 0.926078} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.017550] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Reconfigured VM instance instance-00000050 to attach disk [datastore1] d63fd465-975f-42ce-b1be-ac8a6929bc4d/d63fd465-975f-42ce-b1be-ac8a6929bc4d.vmdk or device None with type streamOptimized {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 923.021286] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-09445a72-f9ea-4100-8b0b-9db8f0884ab8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.028027] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 923.028027] env[63021]: value = "task-1293959" [ 923.028027] env[63021]: _type = "Task" [ 923.028027] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.037989] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293959, 'name': Rename_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.058171] env[63021]: DEBUG nova.network.neutron [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 923.124953] env[63021]: DEBUG oslo_vmware.api [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293958, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.421508} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.125324] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 923.126200] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-880e17f9-4307-4263-93d0-253a5dc41121 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.153499] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] 140184cd-6e4c-425c-8b17-361a1f565e93/140184cd-6e4c-425c-8b17-361a1f565e93.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 923.153860] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f5d9bd9d-94e1-4cb9-9e6f-c62b446e3d20 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.177168] env[63021]: DEBUG oslo_vmware.api [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 923.177168] env[63021]: value = "task-1293960" [ 923.177168] env[63021]: _type = "Task" [ 923.177168] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.186500] env[63021]: DEBUG oslo_vmware.api [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293960, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.257237] env[63021]: DEBUG nova.network.neutron [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Updating instance_info_cache with network_info: [{"id": "f99a201d-df02-499f-85d9-2566ac1dd82b", "address": "fa:16:3e:a0:fd:ff", "network": {"id": "3aaeed35-f452-4144-a259-8f440f8a8bf4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-229941363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b672b1d7d7ad454ebfa3e2fd2421c754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf99a201d-df", "ovs_interfaceid": "f99a201d-df02-499f-85d9-2566ac1dd82b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.371995] env[63021]: DEBUG nova.scheduler.client.report [None req-65a46cd5-0e5c-4e9c-bb88-b7b7e09d89a7 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 923.537554] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293959, 'name': Rename_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.687774] env[63021]: DEBUG oslo_vmware.api [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293960, 'name': ReconfigVM_Task, 'duration_secs': 0.346655} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.688085] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Reconfigured VM instance instance-00000051 to attach disk [datastore1] 140184cd-6e4c-425c-8b17-361a1f565e93/140184cd-6e4c-425c-8b17-361a1f565e93.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 923.688716] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0594ca60-9a7e-407d-9da8-80d416e721ba {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.694854] env[63021]: DEBUG oslo_vmware.api [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 923.694854] env[63021]: value = "task-1293961" [ 923.694854] env[63021]: _type = "Task" [ 923.694854] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.703656] env[63021]: DEBUG oslo_vmware.api [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293961, 'name': Rename_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.760681] env[63021]: DEBUG oslo_concurrency.lockutils [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Releasing lock "refresh_cache-ae103118-bddf-46fe-90b9-98b60952ebba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.761582] env[63021]: DEBUG nova.compute.manager [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Instance network_info: |[{"id": "f99a201d-df02-499f-85d9-2566ac1dd82b", "address": "fa:16:3e:a0:fd:ff", "network": {"id": "3aaeed35-f452-4144-a259-8f440f8a8bf4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-229941363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b672b1d7d7ad454ebfa3e2fd2421c754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf99a201d-df", "ovs_interfaceid": "f99a201d-df02-499f-85d9-2566ac1dd82b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 923.762412] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a0:fd:ff', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ead20342-9afa-435e-a22b-b4a903457712', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f99a201d-df02-499f-85d9-2566ac1dd82b', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 923.770257] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Creating folder: Project (b672b1d7d7ad454ebfa3e2fd2421c754). Parent ref: group-v277447. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 923.770623] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a14c5224-b45f-422f-b324-d5055d24416e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.781085] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Created folder: Project (b672b1d7d7ad454ebfa3e2fd2421c754) in parent group-v277447. [ 923.781311] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Creating folder: Instances. Parent ref: group-v277550. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 923.781591] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1e183c7f-9980-4c39-b7ed-8f3ecea8fc40 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.791330] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Created folder: Instances in parent group-v277550. [ 923.791584] env[63021]: DEBUG oslo.service.loopingcall [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 923.791800] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 923.792040] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c7775075-eb23-47f7-acd6-320e779be7d9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.811903] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 923.811903] env[63021]: value = "task-1293964" [ 923.811903] env[63021]: _type = "Task" [ 923.811903] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.821958] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293964, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.910926] env[63021]: DEBUG nova.compute.manager [req-285d61fd-f87c-4d2d-818a-c005e66c21da req-54d51ce3-f451-4a8a-84d5-3ea4c916563f service nova] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Received event network-changed-f99a201d-df02-499f-85d9-2566ac1dd82b {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 923.912030] env[63021]: DEBUG nova.compute.manager [req-285d61fd-f87c-4d2d-818a-c005e66c21da req-54d51ce3-f451-4a8a-84d5-3ea4c916563f service nova] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Refreshing instance network info cache due to event network-changed-f99a201d-df02-499f-85d9-2566ac1dd82b. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 923.912030] env[63021]: DEBUG oslo_concurrency.lockutils [req-285d61fd-f87c-4d2d-818a-c005e66c21da req-54d51ce3-f451-4a8a-84d5-3ea4c916563f service nova] Acquiring lock "refresh_cache-ae103118-bddf-46fe-90b9-98b60952ebba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.912030] env[63021]: DEBUG oslo_concurrency.lockutils [req-285d61fd-f87c-4d2d-818a-c005e66c21da req-54d51ce3-f451-4a8a-84d5-3ea4c916563f service nova] Acquired lock "refresh_cache-ae103118-bddf-46fe-90b9-98b60952ebba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.912210] env[63021]: DEBUG nova.network.neutron [req-285d61fd-f87c-4d2d-818a-c005e66c21da req-54d51ce3-f451-4a8a-84d5-3ea4c916563f service nova] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Refreshing network info cache for port f99a201d-df02-499f-85d9-2566ac1dd82b {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 924.014087] env[63021]: DEBUG nova.compute.manager [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 924.037476] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293959, 'name': Rename_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.040206] env[63021]: DEBUG nova.virt.hardware [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 924.040547] env[63021]: DEBUG nova.virt.hardware [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 924.040730] env[63021]: DEBUG nova.virt.hardware [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 924.040924] env[63021]: DEBUG nova.virt.hardware [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 924.041084] env[63021]: DEBUG nova.virt.hardware [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 924.041236] env[63021]: DEBUG nova.virt.hardware [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 924.041445] env[63021]: DEBUG nova.virt.hardware [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 924.041663] env[63021]: DEBUG nova.virt.hardware [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 924.041888] env[63021]: DEBUG nova.virt.hardware [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 924.042077] env[63021]: DEBUG nova.virt.hardware [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 924.042261] env[63021]: DEBUG nova.virt.hardware [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 924.043091] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcc49894-0652-4d1c-a773-bbb133e03200 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.051429] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a11e2b84-25a4-495d-a940-d4fd5203bee4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.206237] env[63021]: DEBUG oslo_vmware.api [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293961, 'name': Rename_Task, 'duration_secs': 0.154579} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.206237] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 924.206237] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-77325dda-9c0c-402f-a676-dcdf48c1d998 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.213492] env[63021]: DEBUG oslo_vmware.api [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 924.213492] env[63021]: value = "task-1293965" [ 924.213492] env[63021]: _type = "Task" [ 924.213492] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.222039] env[63021]: DEBUG oslo_vmware.api [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293965, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.321449] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293964, 'name': CreateVM_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.384486] env[63021]: DEBUG oslo_concurrency.lockutils [None req-65a46cd5-0e5c-4e9c-bb88-b7b7e09d89a7 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.394s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.387525] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cb8a5d1d-bec9-4830-b36d-432c88f95588 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.501s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.387789] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cb8a5d1d-bec9-4830-b36d-432c88f95588 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.390047] env[63021]: DEBUG oslo_concurrency.lockutils [None req-95b6e88d-8134-48e8-a7c6-826dc81538bf tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.491s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.390106] env[63021]: DEBUG oslo_concurrency.lockutils [None req-95b6e88d-8134-48e8-a7c6-826dc81538bf tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.391853] env[63021]: DEBUG oslo_concurrency.lockutils [None req-473cf5dd-c87b-43a2-b8eb-5efd58a3afb2 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.176s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.392945] env[63021]: DEBUG oslo_concurrency.lockutils [None req-473cf5dd-c87b-43a2-b8eb-5efd58a3afb2 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.394164] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.735s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.396045] env[63021]: INFO nova.compute.claims [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 924.436919] env[63021]: INFO nova.scheduler.client.report [None req-cb8a5d1d-bec9-4830-b36d-432c88f95588 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Deleted allocations for instance 44dadf8e-a727-448f-887d-6408f7ebfbc6 [ 924.445510] env[63021]: INFO nova.scheduler.client.report [None req-473cf5dd-c87b-43a2-b8eb-5efd58a3afb2 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Deleted allocations for instance 7c6d7839-616d-41f5-a909-f8c626477a5e [ 924.452426] env[63021]: INFO nova.scheduler.client.report [None req-95b6e88d-8134-48e8-a7c6-826dc81538bf tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Deleted allocations for instance 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed [ 924.524016] env[63021]: DEBUG nova.network.neutron [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Successfully updated port: 905466dc-1047-469f-b282-768293b6d154 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 924.540418] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293959, 'name': Rename_Task, 'duration_secs': 1.109097} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.540857] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 924.541055] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-186ff6e1-2fb5-468e-b86e-2ab7d7166723 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.548108] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 924.548108] env[63021]: value = "task-1293966" [ 924.548108] env[63021]: _type = "Task" [ 924.548108] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.558992] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293966, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.685047] env[63021]: DEBUG nova.network.neutron [req-285d61fd-f87c-4d2d-818a-c005e66c21da req-54d51ce3-f451-4a8a-84d5-3ea4c916563f service nova] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Updated VIF entry in instance network info cache for port f99a201d-df02-499f-85d9-2566ac1dd82b. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 924.685460] env[63021]: DEBUG nova.network.neutron [req-285d61fd-f87c-4d2d-818a-c005e66c21da req-54d51ce3-f451-4a8a-84d5-3ea4c916563f service nova] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Updating instance_info_cache with network_info: [{"id": "f99a201d-df02-499f-85d9-2566ac1dd82b", "address": "fa:16:3e:a0:fd:ff", "network": {"id": "3aaeed35-f452-4144-a259-8f440f8a8bf4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-229941363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b672b1d7d7ad454ebfa3e2fd2421c754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf99a201d-df", "ovs_interfaceid": "f99a201d-df02-499f-85d9-2566ac1dd82b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.724426] env[63021]: DEBUG oslo_vmware.api [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293965, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.823654] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293964, 'name': CreateVM_Task, 'duration_secs': 0.511256} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.823919] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 924.824670] env[63021]: DEBUG oslo_concurrency.lockutils [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.824949] env[63021]: DEBUG oslo_concurrency.lockutils [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.825389] env[63021]: DEBUG oslo_concurrency.lockutils [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 924.825663] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f7086a6-0570-484b-9ba1-19f929899686 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.831122] env[63021]: DEBUG oslo_vmware.api [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 924.831122] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52d8bc9c-7dd5-3884-5a54-72ff3f99ab42" [ 924.831122] env[63021]: _type = "Task" [ 924.831122] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.839542] env[63021]: DEBUG oslo_vmware.api [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52d8bc9c-7dd5-3884-5a54-72ff3f99ab42, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.955621] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cb8a5d1d-bec9-4830-b36d-432c88f95588 tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Lock "44dadf8e-a727-448f-887d-6408f7ebfbc6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.039s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.958067] env[63021]: DEBUG oslo_concurrency.lockutils [None req-473cf5dd-c87b-43a2-b8eb-5efd58a3afb2 tempest-ServersAdminNegativeTestJSON-1344052768 tempest-ServersAdminNegativeTestJSON-1344052768-project-member] Lock "7c6d7839-616d-41f5-a909-f8c626477a5e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.773s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.961448] env[63021]: DEBUG oslo_concurrency.lockutils [None req-95b6e88d-8134-48e8-a7c6-826dc81538bf tempest-ListServersNegativeTestJSON-729578415 tempest-ListServersNegativeTestJSON-729578415-project-member] Lock "50a1269e-4d83-4cf7-ba14-3b1afb3c43ed" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.317s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.963754] env[63021]: INFO nova.scheduler.client.report [None req-65a46cd5-0e5c-4e9c-bb88-b7b7e09d89a7 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Deleted allocation for migration b36d8c79-3341-4688-9479-6846fc53dcad [ 925.027288] env[63021]: DEBUG oslo_concurrency.lockutils [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "refresh_cache-418a394b-dbb2-47a2-96a9-a1caa4c4bb51" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.027480] env[63021]: DEBUG oslo_concurrency.lockutils [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquired lock "refresh_cache-418a394b-dbb2-47a2-96a9-a1caa4c4bb51" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.027830] env[63021]: DEBUG nova.network.neutron [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 925.062588] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293966, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.188979] env[63021]: DEBUG oslo_concurrency.lockutils [req-285d61fd-f87c-4d2d-818a-c005e66c21da req-54d51ce3-f451-4a8a-84d5-3ea4c916563f service nova] Releasing lock "refresh_cache-ae103118-bddf-46fe-90b9-98b60952ebba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.225721] env[63021]: DEBUG oslo_vmware.api [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293965, 'name': PowerOnVM_Task, 'duration_secs': 0.546251} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.225721] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 925.225877] env[63021]: INFO nova.compute.manager [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Took 10.35 seconds to spawn the instance on the hypervisor. [ 925.227059] env[63021]: DEBUG nova.compute.manager [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 925.227059] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b682949-3d93-4cd3-b3a2-bb48cd310768 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.343636] env[63021]: DEBUG oslo_vmware.api [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52d8bc9c-7dd5-3884-5a54-72ff3f99ab42, 'name': SearchDatastore_Task, 'duration_secs': 0.027992} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.343993] env[63021]: DEBUG oslo_concurrency.lockutils [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.344245] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 925.344488] env[63021]: DEBUG oslo_concurrency.lockutils [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.344818] env[63021]: DEBUG oslo_concurrency.lockutils [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.344818] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 925.345098] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6a830358-c158-4283-beed-67d19d26b38f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.356882] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 925.357060] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 925.357877] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7aac06dc-199c-4131-93b0-9273a3670299 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.363899] env[63021]: DEBUG oslo_vmware.api [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 925.363899] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]520cd674-7891-3abc-1006-bee69e365d9d" [ 925.363899] env[63021]: _type = "Task" [ 925.363899] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.372721] env[63021]: DEBUG oslo_vmware.api [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]520cd674-7891-3abc-1006-bee69e365d9d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.470861] env[63021]: DEBUG oslo_concurrency.lockutils [None req-65a46cd5-0e5c-4e9c-bb88-b7b7e09d89a7 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "71909213-258d-45f5-9a3e-e473e8ce8aa2" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 18.288s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.562594] env[63021]: DEBUG oslo_vmware.api [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293966, 'name': PowerOnVM_Task, 'duration_secs': 0.926908} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.566867] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 925.566867] env[63021]: INFO nova.compute.manager [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Took 19.27 seconds to spawn the instance on the hypervisor. [ 925.566867] env[63021]: DEBUG nova.compute.manager [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 925.566867] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8894a12-70a3-4000-9743-9cf3cb115ef2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.577368] env[63021]: DEBUG nova.network.neutron [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 925.689401] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb9f52e6-665e-4300-812d-b1cd07052292 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.699930] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02f752de-a2a4-42c9-a7b2-6dc8b078c9f6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.741864] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae1b7e8a-7ae9-4941-9a8f-a510a5d28aa8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.751205] env[63021]: INFO nova.compute.manager [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Took 21.45 seconds to build instance. [ 925.758207] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1333dac-02d0-42fb-98cf-be51da0814bb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.777182] env[63021]: DEBUG nova.compute.provider_tree [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 925.877549] env[63021]: DEBUG oslo_vmware.api [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]520cd674-7891-3abc-1006-bee69e365d9d, 'name': SearchDatastore_Task, 'duration_secs': 0.012501} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.878466] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bca5729c-14d9-47d1-ad11-77c79309aa82 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.884837] env[63021]: DEBUG oslo_vmware.api [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 925.884837] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52ad238a-b0b7-8d27-e308-e619fd4492fe" [ 925.884837] env[63021]: _type = "Task" [ 925.884837] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.892925] env[63021]: DEBUG oslo_vmware.api [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52ad238a-b0b7-8d27-e308-e619fd4492fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.934452] env[63021]: DEBUG nova.network.neutron [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Updating instance_info_cache with network_info: [{"id": "905466dc-1047-469f-b282-768293b6d154", "address": "fa:16:3e:db:7b:17", "network": {"id": "3aaeed35-f452-4144-a259-8f440f8a8bf4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-229941363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b672b1d7d7ad454ebfa3e2fd2421c754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap905466dc-10", "ovs_interfaceid": "905466dc-1047-469f-b282-768293b6d154", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.982400] env[63021]: DEBUG nova.compute.manager [req-33000419-1429-4dfe-89e6-8c78d2993011 req-81c5608e-5097-4c1e-bfe7-b0b41fa3d027 service nova] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Received event network-vif-plugged-905466dc-1047-469f-b282-768293b6d154 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 925.982633] env[63021]: DEBUG oslo_concurrency.lockutils [req-33000419-1429-4dfe-89e6-8c78d2993011 req-81c5608e-5097-4c1e-bfe7-b0b41fa3d027 service nova] Acquiring lock "418a394b-dbb2-47a2-96a9-a1caa4c4bb51-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.982962] env[63021]: DEBUG oslo_concurrency.lockutils [req-33000419-1429-4dfe-89e6-8c78d2993011 req-81c5608e-5097-4c1e-bfe7-b0b41fa3d027 service nova] Lock "418a394b-dbb2-47a2-96a9-a1caa4c4bb51-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.983104] env[63021]: DEBUG oslo_concurrency.lockutils [req-33000419-1429-4dfe-89e6-8c78d2993011 req-81c5608e-5097-4c1e-bfe7-b0b41fa3d027 service nova] Lock "418a394b-dbb2-47a2-96a9-a1caa4c4bb51-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.983327] env[63021]: DEBUG nova.compute.manager [req-33000419-1429-4dfe-89e6-8c78d2993011 req-81c5608e-5097-4c1e-bfe7-b0b41fa3d027 service nova] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] No waiting events found dispatching network-vif-plugged-905466dc-1047-469f-b282-768293b6d154 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 925.983535] env[63021]: WARNING nova.compute.manager [req-33000419-1429-4dfe-89e6-8c78d2993011 req-81c5608e-5097-4c1e-bfe7-b0b41fa3d027 service nova] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Received unexpected event network-vif-plugged-905466dc-1047-469f-b282-768293b6d154 for instance with vm_state building and task_state spawning. [ 925.983752] env[63021]: DEBUG nova.compute.manager [req-33000419-1429-4dfe-89e6-8c78d2993011 req-81c5608e-5097-4c1e-bfe7-b0b41fa3d027 service nova] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Received event network-changed-905466dc-1047-469f-b282-768293b6d154 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 925.983926] env[63021]: DEBUG nova.compute.manager [req-33000419-1429-4dfe-89e6-8c78d2993011 req-81c5608e-5097-4c1e-bfe7-b0b41fa3d027 service nova] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Refreshing instance network info cache due to event network-changed-905466dc-1047-469f-b282-768293b6d154. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 925.984188] env[63021]: DEBUG oslo_concurrency.lockutils [req-33000419-1429-4dfe-89e6-8c78d2993011 req-81c5608e-5097-4c1e-bfe7-b0b41fa3d027 service nova] Acquiring lock "refresh_cache-418a394b-dbb2-47a2-96a9-a1caa4c4bb51" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.091187] env[63021]: INFO nova.compute.manager [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Took 31.40 seconds to build instance. [ 926.254523] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6f5c0636-800e-401e-83af-055cfd4b69ef tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "140184cd-6e4c-425c-8b17-361a1f565e93" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.970s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.278852] env[63021]: DEBUG nova.scheduler.client.report [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 926.399114] env[63021]: DEBUG oslo_vmware.api [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52ad238a-b0b7-8d27-e308-e619fd4492fe, 'name': SearchDatastore_Task, 'duration_secs': 0.021776} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.399503] env[63021]: DEBUG oslo_concurrency.lockutils [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.399890] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] ae103118-bddf-46fe-90b9-98b60952ebba/ae103118-bddf-46fe-90b9-98b60952ebba.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 926.400410] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-929003fb-d958-4247-a20b-bb23d4f4d9ee {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.407835] env[63021]: DEBUG oslo_vmware.api [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 926.407835] env[63021]: value = "task-1293967" [ 926.407835] env[63021]: _type = "Task" [ 926.407835] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.417267] env[63021]: DEBUG oslo_vmware.api [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1293967, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.438743] env[63021]: DEBUG oslo_concurrency.lockutils [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Releasing lock "refresh_cache-418a394b-dbb2-47a2-96a9-a1caa4c4bb51" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.439081] env[63021]: DEBUG nova.compute.manager [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Instance network_info: |[{"id": "905466dc-1047-469f-b282-768293b6d154", "address": "fa:16:3e:db:7b:17", "network": {"id": "3aaeed35-f452-4144-a259-8f440f8a8bf4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-229941363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b672b1d7d7ad454ebfa3e2fd2421c754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap905466dc-10", "ovs_interfaceid": "905466dc-1047-469f-b282-768293b6d154", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 926.439665] env[63021]: DEBUG oslo_concurrency.lockutils [req-33000419-1429-4dfe-89e6-8c78d2993011 req-81c5608e-5097-4c1e-bfe7-b0b41fa3d027 service nova] Acquired lock "refresh_cache-418a394b-dbb2-47a2-96a9-a1caa4c4bb51" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.439854] env[63021]: DEBUG nova.network.neutron [req-33000419-1429-4dfe-89e6-8c78d2993011 req-81c5608e-5097-4c1e-bfe7-b0b41fa3d027 service nova] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Refreshing network info cache for port 905466dc-1047-469f-b282-768293b6d154 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 926.441092] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:db:7b:17', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ead20342-9afa-435e-a22b-b4a903457712', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '905466dc-1047-469f-b282-768293b6d154', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 926.451207] env[63021]: DEBUG oslo.service.loopingcall [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 926.455092] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 926.455608] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ce75051c-3f6d-4a71-b94b-673775078c25 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.482670] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 926.482670] env[63021]: value = "task-1293968" [ 926.482670] env[63021]: _type = "Task" [ 926.482670] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.494288] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293968, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.594613] env[63021]: DEBUG oslo_concurrency.lockutils [None req-74365e89-5a28-4c62-b60d-6ad3eda9dcbe tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "d63fd465-975f-42ce-b1be-ac8a6929bc4d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.909s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.786747] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.393s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.787354] env[63021]: DEBUG nova.compute.manager [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 926.796271] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.420s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.796271] env[63021]: DEBUG nova.objects.instance [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Lazy-loading 'resources' on Instance uuid 1b4871cc-3dac-4683-8fce-93c4d9e4407d {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.821030] env[63021]: DEBUG oslo_vmware.rw_handles [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a434df-1d61-e144-a3f8-ffe1b46c8c86/disk-0.vmdk. {{(pid=63021) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 926.822657] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a45babc4-1627-4511-81a4-ad47b09c958f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.829801] env[63021]: DEBUG oslo_vmware.rw_handles [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a434df-1d61-e144-a3f8-ffe1b46c8c86/disk-0.vmdk is in state: ready. {{(pid=63021) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 926.830917] env[63021]: ERROR oslo_vmware.rw_handles [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a434df-1d61-e144-a3f8-ffe1b46c8c86/disk-0.vmdk due to incomplete transfer. [ 926.830917] env[63021]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-89342b7f-6b1b-41c0-9ccf-39f82e37ce62 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.842582] env[63021]: DEBUG oslo_vmware.rw_handles [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a434df-1d61-e144-a3f8-ffe1b46c8c86/disk-0.vmdk. {{(pid=63021) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 926.842922] env[63021]: DEBUG nova.virt.vmwareapi.images [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Uploaded image d9df5b42-2840-4afc-82fa-268aed408575 to the Glance image server {{(pid=63021) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 926.845669] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Destroying the VM {{(pid=63021) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 926.847332] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-2b12f471-1455-4a24-9db5-dec515f9f954 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.859839] env[63021]: DEBUG oslo_vmware.api [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Waiting for the task: (returnval){ [ 926.859839] env[63021]: value = "task-1293969" [ 926.859839] env[63021]: _type = "Task" [ 926.859839] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.873828] env[63021]: DEBUG oslo_vmware.api [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1293969, 'name': Destroy_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.920118] env[63021]: DEBUG oslo_vmware.api [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1293967, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.935031] env[63021]: DEBUG oslo_concurrency.lockutils [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "d63fd465-975f-42ce-b1be-ac8a6929bc4d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.935449] env[63021]: DEBUG oslo_concurrency.lockutils [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "d63fd465-975f-42ce-b1be-ac8a6929bc4d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.935695] env[63021]: DEBUG oslo_concurrency.lockutils [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "d63fd465-975f-42ce-b1be-ac8a6929bc4d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.935886] env[63021]: DEBUG oslo_concurrency.lockutils [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "d63fd465-975f-42ce-b1be-ac8a6929bc4d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.936071] env[63021]: DEBUG oslo_concurrency.lockutils [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "d63fd465-975f-42ce-b1be-ac8a6929bc4d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.942395] env[63021]: INFO nova.compute.manager [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Terminating instance [ 926.945805] env[63021]: DEBUG nova.compute.manager [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 926.946051] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 926.947034] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f74f2693-7414-46bc-8752-e9178800719e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.957118] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 926.957775] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4436c713-20c4-408a-b89e-913ef810b6f7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.967683] env[63021]: DEBUG oslo_vmware.api [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 926.967683] env[63021]: value = "task-1293970" [ 926.967683] env[63021]: _type = "Task" [ 926.967683] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.979284] env[63021]: DEBUG oslo_vmware.api [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293970, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.993254] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293968, 'name': CreateVM_Task, 'duration_secs': 0.398877} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.993449] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 926.994182] env[63021]: DEBUG oslo_concurrency.lockutils [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.994366] env[63021]: DEBUG oslo_concurrency.lockutils [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.994735] env[63021]: DEBUG oslo_concurrency.lockutils [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 926.995017] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0986b8e-da26-449e-9b7e-9cc977efc8be {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.000539] env[63021]: DEBUG oslo_vmware.api [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 927.000539] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5213a40c-543e-b7f1-768f-6f619015ab42" [ 927.000539] env[63021]: _type = "Task" [ 927.000539] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.013677] env[63021]: DEBUG oslo_vmware.api [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5213a40c-543e-b7f1-768f-6f619015ab42, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.173509] env[63021]: DEBUG oslo_concurrency.lockutils [None req-24e72953-af49-46ed-99ac-23f23801a0d0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "140184cd-6e4c-425c-8b17-361a1f565e93" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.173801] env[63021]: DEBUG oslo_concurrency.lockutils [None req-24e72953-af49-46ed-99ac-23f23801a0d0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "140184cd-6e4c-425c-8b17-361a1f565e93" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.296459] env[63021]: DEBUG nova.network.neutron [req-33000419-1429-4dfe-89e6-8c78d2993011 req-81c5608e-5097-4c1e-bfe7-b0b41fa3d027 service nova] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Updated VIF entry in instance network info cache for port 905466dc-1047-469f-b282-768293b6d154. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 927.299037] env[63021]: DEBUG nova.network.neutron [req-33000419-1429-4dfe-89e6-8c78d2993011 req-81c5608e-5097-4c1e-bfe7-b0b41fa3d027 service nova] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Updating instance_info_cache with network_info: [{"id": "905466dc-1047-469f-b282-768293b6d154", "address": "fa:16:3e:db:7b:17", "network": {"id": "3aaeed35-f452-4144-a259-8f440f8a8bf4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-229941363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b672b1d7d7ad454ebfa3e2fd2421c754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap905466dc-10", "ovs_interfaceid": "905466dc-1047-469f-b282-768293b6d154", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.299604] env[63021]: DEBUG nova.compute.utils [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 927.311337] env[63021]: DEBUG nova.compute.manager [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 927.311337] env[63021]: DEBUG nova.network.neutron [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 927.378435] env[63021]: DEBUG oslo_vmware.api [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1293969, 'name': Destroy_Task} progress is 33%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.386783] env[63021]: DEBUG nova.policy [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6ea19453dbdb4251a93dde2576aeff64', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '58693d1e991249f28914a39761cc2627', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 927.419065] env[63021]: DEBUG oslo_vmware.api [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1293967, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.636273} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.419363] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] ae103118-bddf-46fe-90b9-98b60952ebba/ae103118-bddf-46fe-90b9-98b60952ebba.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 927.419590] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 927.419869] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c1fb8312-1f43-484b-8c3a-8680d480435f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.428533] env[63021]: DEBUG oslo_vmware.api [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 927.428533] env[63021]: value = "task-1293971" [ 927.428533] env[63021]: _type = "Task" [ 927.428533] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.438018] env[63021]: DEBUG oslo_vmware.api [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1293971, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.476947] env[63021]: DEBUG oslo_vmware.api [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293970, 'name': PowerOffVM_Task, 'duration_secs': 0.251596} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.477242] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 927.477440] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 927.477718] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bcbf687d-9b7c-4ec5-af89-54e4c9f159a0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.511245] env[63021]: DEBUG oslo_vmware.api [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5213a40c-543e-b7f1-768f-6f619015ab42, 'name': SearchDatastore_Task, 'duration_secs': 0.028951} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.517053] env[63021]: DEBUG oslo_concurrency.lockutils [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.517053] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 927.517053] env[63021]: DEBUG oslo_concurrency.lockutils [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.517053] env[63021]: DEBUG oslo_concurrency.lockutils [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.517053] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 927.517053] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5c917c63-79f4-4df3-bcac-66086d8d6221 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.527168] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 927.527614] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 927.530908] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af860802-b654-49eb-a712-abb2e76dbca8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.536866] env[63021]: DEBUG oslo_vmware.api [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 927.536866] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52d10474-4dd7-2900-1dc6-4710d70e811c" [ 927.536866] env[63021]: _type = "Task" [ 927.536866] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.546624] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 927.546624] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 927.546624] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Deleting the datastore file [datastore1] d63fd465-975f-42ce-b1be-ac8a6929bc4d {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 927.550758] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-70aace65-4e65-4aa4-bef7-8ae55018f32e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.552836] env[63021]: DEBUG oslo_vmware.api [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52d10474-4dd7-2900-1dc6-4710d70e811c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.557285] env[63021]: DEBUG oslo_vmware.api [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 927.557285] env[63021]: value = "task-1293973" [ 927.557285] env[63021]: _type = "Task" [ 927.557285] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.570995] env[63021]: DEBUG oslo_vmware.api [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293973, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.602019] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49c183f2-0305-4eec-86b2-b053356e1630 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.608175] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1d71513-32cb-471b-9e48-1e2383ecd1ff {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.644336] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9956aca3-e1c6-4db4-a4b2-6e59270aecb6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.657462] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34e49bd5-61d3-400a-ba17-814bb895046e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.674206] env[63021]: DEBUG nova.compute.provider_tree [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 927.678021] env[63021]: DEBUG nova.compute.utils [None req-24e72953-af49-46ed-99ac-23f23801a0d0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 927.809939] env[63021]: DEBUG oslo_concurrency.lockutils [req-33000419-1429-4dfe-89e6-8c78d2993011 req-81c5608e-5097-4c1e-bfe7-b0b41fa3d027 service nova] Releasing lock "refresh_cache-418a394b-dbb2-47a2-96a9-a1caa4c4bb51" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.810626] env[63021]: DEBUG nova.compute.manager [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 927.875810] env[63021]: DEBUG oslo_vmware.api [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1293969, 'name': Destroy_Task, 'duration_secs': 0.553008} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.875810] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Destroyed the VM [ 927.875810] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Deleting Snapshot of the VM instance {{(pid=63021) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 927.875810] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-83beffd8-0a2c-4167-a718-1d751ba259da {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.880421] env[63021]: DEBUG oslo_vmware.api [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Waiting for the task: (returnval){ [ 927.880421] env[63021]: value = "task-1293974" [ 927.880421] env[63021]: _type = "Task" [ 927.880421] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.889824] env[63021]: DEBUG oslo_vmware.api [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1293974, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.941058] env[63021]: DEBUG oslo_vmware.api [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1293971, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.129158} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.941058] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 927.941058] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72100532-6f6b-4564-9e0c-35b369c645cf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.968029] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] ae103118-bddf-46fe-90b9-98b60952ebba/ae103118-bddf-46fe-90b9-98b60952ebba.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 927.968029] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-888bd1a9-ecf6-4ff8-a73d-4fea9e9e3c8a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.989133] env[63021]: DEBUG oslo_vmware.api [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 927.989133] env[63021]: value = "task-1293975" [ 927.989133] env[63021]: _type = "Task" [ 927.989133] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.997667] env[63021]: DEBUG oslo_vmware.api [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1293975, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.048061] env[63021]: DEBUG oslo_vmware.api [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52d10474-4dd7-2900-1dc6-4710d70e811c, 'name': SearchDatastore_Task, 'duration_secs': 0.012126} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.048830] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55f65bb9-3b9c-4346-99ed-ebb78626b46c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.054901] env[63021]: DEBUG oslo_vmware.api [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 928.054901] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52306131-fd94-a0b0-852f-56f75e895d5a" [ 928.054901] env[63021]: _type = "Task" [ 928.054901] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.067024] env[63021]: DEBUG oslo_vmware.api [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52306131-fd94-a0b0-852f-56f75e895d5a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.070134] env[63021]: DEBUG oslo_vmware.api [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1293973, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.238926} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.070346] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 928.070533] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 928.070741] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 928.070918] env[63021]: INFO nova.compute.manager [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Took 1.12 seconds to destroy the instance on the hypervisor. [ 928.071176] env[63021]: DEBUG oslo.service.loopingcall [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 928.071374] env[63021]: DEBUG nova.compute.manager [-] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 928.071471] env[63021]: DEBUG nova.network.neutron [-] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 928.151577] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d0020b0d-18e8-4912-9dc0-605a5e814d35 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "71909213-258d-45f5-9a3e-e473e8ce8aa2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.151878] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d0020b0d-18e8-4912-9dc0-605a5e814d35 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "71909213-258d-45f5-9a3e-e473e8ce8aa2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.152112] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d0020b0d-18e8-4912-9dc0-605a5e814d35 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "71909213-258d-45f5-9a3e-e473e8ce8aa2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.152306] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d0020b0d-18e8-4912-9dc0-605a5e814d35 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "71909213-258d-45f5-9a3e-e473e8ce8aa2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.152479] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d0020b0d-18e8-4912-9dc0-605a5e814d35 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "71909213-258d-45f5-9a3e-e473e8ce8aa2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.155093] env[63021]: INFO nova.compute.manager [None req-d0020b0d-18e8-4912-9dc0-605a5e814d35 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Terminating instance [ 928.158180] env[63021]: DEBUG nova.network.neutron [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Successfully created port: eab7f4fe-4832-40e4-aed2-0e931d91a4d7 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 928.160709] env[63021]: DEBUG nova.compute.manager [None req-d0020b0d-18e8-4912-9dc0-605a5e814d35 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 928.160907] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d0020b0d-18e8-4912-9dc0-605a5e814d35 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 928.162055] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc523d70-89dc-4bbc-bd82-d394622f5195 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.170274] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0020b0d-18e8-4912-9dc0-605a5e814d35 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 928.170520] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-266a8a4a-7776-40ba-b341-ca4beb756956 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.176185] env[63021]: DEBUG oslo_vmware.api [None req-d0020b0d-18e8-4912-9dc0-605a5e814d35 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 928.176185] env[63021]: value = "task-1293976" [ 928.176185] env[63021]: _type = "Task" [ 928.176185] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.177037] env[63021]: DEBUG nova.scheduler.client.report [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 928.187208] env[63021]: DEBUG oslo_concurrency.lockutils [None req-24e72953-af49-46ed-99ac-23f23801a0d0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "140184cd-6e4c-425c-8b17-361a1f565e93" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.011s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.190166] env[63021]: DEBUG oslo_vmware.api [None req-d0020b0d-18e8-4912-9dc0-605a5e814d35 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1293976, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.395067] env[63021]: DEBUG oslo_vmware.api [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1293974, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.498990] env[63021]: DEBUG oslo_vmware.api [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1293975, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.566896] env[63021]: DEBUG oslo_vmware.api [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52306131-fd94-a0b0-852f-56f75e895d5a, 'name': SearchDatastore_Task, 'duration_secs': 0.013553} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.567210] env[63021]: DEBUG oslo_concurrency.lockutils [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.567550] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 418a394b-dbb2-47a2-96a9-a1caa4c4bb51/418a394b-dbb2-47a2-96a9-a1caa4c4bb51.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 928.567878] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-97b11f82-9cc2-4263-8ee7-e6234e6a88bc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.573762] env[63021]: DEBUG oslo_vmware.api [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 928.573762] env[63021]: value = "task-1293977" [ 928.573762] env[63021]: _type = "Task" [ 928.573762] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.583510] env[63021]: DEBUG oslo_vmware.api [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1293977, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.683534] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.887s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.688908] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 12.027s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.688908] env[63021]: DEBUG nova.objects.instance [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63021) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 928.700619] env[63021]: DEBUG oslo_vmware.api [None req-d0020b0d-18e8-4912-9dc0-605a5e814d35 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1293976, 'name': PowerOffVM_Task, 'duration_secs': 0.384499} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.700619] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0020b0d-18e8-4912-9dc0-605a5e814d35 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 928.700619] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d0020b0d-18e8-4912-9dc0-605a5e814d35 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 928.700619] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1d4aae38-ea2c-407a-80d3-c9880e25ea61 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.717124] env[63021]: INFO nova.scheduler.client.report [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Deleted allocations for instance 1b4871cc-3dac-4683-8fce-93c4d9e4407d [ 928.741608] env[63021]: DEBUG nova.compute.manager [req-9808080e-68b0-4ef1-8b64-e7ceea9de079 req-7f7a8ac8-7b54-470e-aad1-3aea4d57b571 service nova] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Received event network-vif-deleted-eb6c1950-a21a-4873-9c10-53065bf84e01 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 928.741807] env[63021]: INFO nova.compute.manager [req-9808080e-68b0-4ef1-8b64-e7ceea9de079 req-7f7a8ac8-7b54-470e-aad1-3aea4d57b571 service nova] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Neutron deleted interface eb6c1950-a21a-4873-9c10-53065bf84e01; detaching it from the instance and deleting it from the info cache [ 928.741980] env[63021]: DEBUG nova.network.neutron [req-9808080e-68b0-4ef1-8b64-e7ceea9de079 req-7f7a8ac8-7b54-470e-aad1-3aea4d57b571 service nova] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.769933] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d0020b0d-18e8-4912-9dc0-605a5e814d35 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 928.770467] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d0020b0d-18e8-4912-9dc0-605a5e814d35 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Deleting contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 928.771117] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0020b0d-18e8-4912-9dc0-605a5e814d35 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Deleting the datastore file [datastore2] 71909213-258d-45f5-9a3e-e473e8ce8aa2 {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 928.771645] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-09d324e9-3bcd-4c7a-99b7-abad8015d23d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.784438] env[63021]: DEBUG oslo_vmware.api [None req-d0020b0d-18e8-4912-9dc0-605a5e814d35 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 928.784438] env[63021]: value = "task-1293979" [ 928.784438] env[63021]: _type = "Task" [ 928.784438] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.800087] env[63021]: DEBUG oslo_vmware.api [None req-d0020b0d-18e8-4912-9dc0-605a5e814d35 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1293979, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.821728] env[63021]: DEBUG nova.compute.manager [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 928.848976] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "207c8197-6191-4837-8ead-8c24aa5b35bc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.849577] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "207c8197-6191-4837-8ead-8c24aa5b35bc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.857543] env[63021]: DEBUG nova.virt.hardware [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 928.861033] env[63021]: DEBUG nova.virt.hardware [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 928.861033] env[63021]: DEBUG nova.virt.hardware [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 928.861033] env[63021]: DEBUG nova.virt.hardware [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 928.861033] env[63021]: DEBUG nova.virt.hardware [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 928.861033] env[63021]: DEBUG nova.virt.hardware [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 928.861033] env[63021]: DEBUG nova.virt.hardware [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 928.861033] env[63021]: DEBUG nova.virt.hardware [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 928.861033] env[63021]: DEBUG nova.virt.hardware [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 928.861033] env[63021]: DEBUG nova.virt.hardware [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 928.861033] env[63021]: DEBUG nova.virt.hardware [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 928.861033] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f48e886f-c809-4f55-8dfd-9c353e849802 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.874905] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-893f0630-26bc-43cd-aba8-730cdd14f779 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.898149] env[63021]: DEBUG oslo_vmware.api [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1293974, 'name': RemoveSnapshot_Task, 'duration_secs': 0.526793} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.898504] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Deleted Snapshot of the VM instance {{(pid=63021) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 928.898833] env[63021]: DEBUG nova.compute.manager [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 928.899684] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-619df82e-49a1-4956-a454-8ed5242592ae {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.002622] env[63021]: DEBUG oslo_vmware.api [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1293975, 'name': ReconfigVM_Task, 'duration_secs': 0.668038} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.003063] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Reconfigured VM instance instance-00000052 to attach disk [datastore1] ae103118-bddf-46fe-90b9-98b60952ebba/ae103118-bddf-46fe-90b9-98b60952ebba.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 929.003920] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-562d2c15-9dba-4153-a468-73c1cbb0d0f2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.011558] env[63021]: DEBUG oslo_vmware.api [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 929.011558] env[63021]: value = "task-1293980" [ 929.011558] env[63021]: _type = "Task" [ 929.011558] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.024235] env[63021]: DEBUG oslo_vmware.api [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1293980, 'name': Rename_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.086721] env[63021]: DEBUG oslo_vmware.api [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1293977, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.215890] env[63021]: DEBUG nova.network.neutron [-] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.224879] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c76b65bc-b9e7-4df0-a480-0cb18315890c tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Lock "1b4871cc-3dac-4683-8fce-93c4d9e4407d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.143s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.246336] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1ec3e89a-0d10-4142-b1ca-544c726c163c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.258172] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-799145ca-5839-457f-b6d0-42c1499cf8a3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.292051] env[63021]: DEBUG oslo_concurrency.lockutils [None req-24e72953-af49-46ed-99ac-23f23801a0d0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "140184cd-6e4c-425c-8b17-361a1f565e93" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.292185] env[63021]: DEBUG oslo_concurrency.lockutils [None req-24e72953-af49-46ed-99ac-23f23801a0d0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "140184cd-6e4c-425c-8b17-361a1f565e93" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.292324] env[63021]: INFO nova.compute.manager [None req-24e72953-af49-46ed-99ac-23f23801a0d0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Attaching volume 460ebecc-27ec-4047-99f6-a80a6e32f71d to /dev/sdb [ 929.294202] env[63021]: DEBUG nova.compute.manager [req-9808080e-68b0-4ef1-8b64-e7ceea9de079 req-7f7a8ac8-7b54-470e-aad1-3aea4d57b571 service nova] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Detach interface failed, port_id=eb6c1950-a21a-4873-9c10-53065bf84e01, reason: Instance d63fd465-975f-42ce-b1be-ac8a6929bc4d could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 929.305068] env[63021]: DEBUG oslo_vmware.api [None req-d0020b0d-18e8-4912-9dc0-605a5e814d35 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1293979, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.325334} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.305207] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0020b0d-18e8-4912-9dc0-605a5e814d35 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 929.305431] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d0020b0d-18e8-4912-9dc0-605a5e814d35 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Deleted contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 929.305627] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d0020b0d-18e8-4912-9dc0-605a5e814d35 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 929.305835] env[63021]: INFO nova.compute.manager [None req-d0020b0d-18e8-4912-9dc0-605a5e814d35 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Took 1.14 seconds to destroy the instance on the hypervisor. [ 929.306044] env[63021]: DEBUG oslo.service.loopingcall [None req-d0020b0d-18e8-4912-9dc0-605a5e814d35 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 929.306248] env[63021]: DEBUG nova.compute.manager [-] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 929.306349] env[63021]: DEBUG nova.network.neutron [-] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 929.332421] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3987e7e-0ec0-430b-954e-bbf8036bedb2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.339099] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d24c366e-d7da-463e-9ac8-9ea7ea54017e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.352417] env[63021]: DEBUG nova.compute.manager [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 929.355123] env[63021]: DEBUG nova.virt.block_device [None req-24e72953-af49-46ed-99ac-23f23801a0d0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Updating existing volume attachment record: fa74bf0e-0e5c-4a67-b516-46cac42ab8ec {{(pid=63021) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 929.415481] env[63021]: INFO nova.compute.manager [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Shelve offloading [ 929.420588] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 929.421181] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a9fcded1-087f-489f-b04d-61a9dd7e9ef0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.427959] env[63021]: DEBUG oslo_vmware.api [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Waiting for the task: (returnval){ [ 929.427959] env[63021]: value = "task-1293981" [ 929.427959] env[63021]: _type = "Task" [ 929.427959] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.437462] env[63021]: DEBUG oslo_vmware.api [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1293981, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.527406] env[63021]: DEBUG oslo_vmware.api [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1293980, 'name': Rename_Task, 'duration_secs': 0.171624} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.527566] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 929.527837] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-89bb4db7-f3be-44a6-80fb-8e0e1640e2c7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.534096] env[63021]: DEBUG oslo_vmware.api [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 929.534096] env[63021]: value = "task-1293983" [ 929.534096] env[63021]: _type = "Task" [ 929.534096] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.549132] env[63021]: DEBUG oslo_vmware.api [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1293983, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.588797] env[63021]: DEBUG oslo_vmware.api [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1293977, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.702187} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.588797] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 418a394b-dbb2-47a2-96a9-a1caa4c4bb51/418a394b-dbb2-47a2-96a9-a1caa4c4bb51.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 929.588909] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 929.589215] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c009d262-d479-4e7e-8ec4-404d05e1876c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.598959] env[63021]: DEBUG oslo_vmware.api [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 929.598959] env[63021]: value = "task-1293984" [ 929.598959] env[63021]: _type = "Task" [ 929.598959] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.610796] env[63021]: DEBUG oslo_vmware.api [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1293984, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.702263] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e9bfeb87-cd93-4640-bff6-04e6a01e2943 tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.703590] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.821s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.704413] env[63021]: DEBUG nova.objects.instance [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Lazy-loading 'resources' on Instance uuid 4d712cc1-6202-407e-8767-c620a1ce0bba {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 929.718226] env[63021]: INFO nova.compute.manager [-] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Took 1.65 seconds to deallocate network for instance. [ 929.886417] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.944260] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] VM already powered off {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 929.944532] env[63021]: DEBUG nova.compute.manager [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 929.945514] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df784a06-25a4-4da3-93cf-86b729b0227f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.952537] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Acquiring lock "refresh_cache-45c9e2be-eb19-4b83-b280-c9eeaddfccba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.952738] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Acquired lock "refresh_cache-45c9e2be-eb19-4b83-b280-c9eeaddfccba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.952896] env[63021]: DEBUG nova.network.neutron [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 930.023754] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquiring lock "df0b6974-7f78-4b73-8583-d09754fbacea" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.024447] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Lock "df0b6974-7f78-4b73-8583-d09754fbacea" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.024706] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquiring lock "df0b6974-7f78-4b73-8583-d09754fbacea-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.024900] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Lock "df0b6974-7f78-4b73-8583-d09754fbacea-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.025101] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Lock "df0b6974-7f78-4b73-8583-d09754fbacea-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.033737] env[63021]: INFO nova.compute.manager [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Terminating instance [ 930.036255] env[63021]: DEBUG nova.compute.manager [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 930.036488] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 930.040656] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48fd7bc8-6350-476a-bb94-c46b4721a1de {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.054242] env[63021]: DEBUG oslo_vmware.api [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1293983, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.056673] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 930.056970] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-98569d0b-ba9b-4fa1-9b2a-2c1d1bc34594 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.063535] env[63021]: DEBUG oslo_vmware.api [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 930.063535] env[63021]: value = "task-1293987" [ 930.063535] env[63021]: _type = "Task" [ 930.063535] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.071957] env[63021]: DEBUG oslo_vmware.api [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293987, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.097474] env[63021]: DEBUG nova.network.neutron [-] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.108944] env[63021]: DEBUG oslo_vmware.api [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1293984, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066633} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.108944] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 930.109648] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4a9c0aa-c6b0-45aa-9dc3-3d850f057eaf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.134915] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] 418a394b-dbb2-47a2-96a9-a1caa4c4bb51/418a394b-dbb2-47a2-96a9-a1caa4c4bb51.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 930.134915] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f8726526-74ad-422a-8134-a08e7aa56692 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.151584] env[63021]: DEBUG nova.network.neutron [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Successfully updated port: eab7f4fe-4832-40e4-aed2-0e931d91a4d7 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 930.158908] env[63021]: DEBUG oslo_vmware.api [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 930.158908] env[63021]: value = "task-1293988" [ 930.158908] env[63021]: _type = "Task" [ 930.158908] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.171318] env[63021]: DEBUG oslo_vmware.api [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1293988, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.229026] env[63021]: DEBUG oslo_concurrency.lockutils [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.442974] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94f9e94f-7357-47f6-86a9-81d34b0e96bd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.451894] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75d627d6-d04a-486d-af3f-5c6f83abd5a7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.486720] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ec21467-be49-4383-81c1-8dcbfdea74b8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.496116] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-450b7049-9efd-423a-bfc4-983f070eba7c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.510534] env[63021]: DEBUG nova.compute.provider_tree [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 930.552020] env[63021]: DEBUG oslo_vmware.api [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1293983, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.572924] env[63021]: DEBUG oslo_vmware.api [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293987, 'name': PowerOffVM_Task, 'duration_secs': 0.240487} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.573230] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 930.573425] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 930.573692] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-648f06af-ef20-422c-948e-5069da8316ee {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.604778] env[63021]: INFO nova.compute.manager [-] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Took 1.30 seconds to deallocate network for instance. [ 930.653418] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "refresh_cache-2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.653482] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquired lock "refresh_cache-2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.653676] env[63021]: DEBUG nova.network.neutron [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 930.671949] env[63021]: DEBUG oslo_vmware.api [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1293988, 'name': ReconfigVM_Task, 'duration_secs': 0.500088} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.672327] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Reconfigured VM instance instance-00000053 to attach disk [datastore1] 418a394b-dbb2-47a2-96a9-a1caa4c4bb51/418a394b-dbb2-47a2-96a9-a1caa4c4bb51.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 930.673069] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-497fd8ec-e583-40c3-b033-9210d60e3634 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.679675] env[63021]: DEBUG oslo_vmware.api [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 930.679675] env[63021]: value = "task-1293990" [ 930.679675] env[63021]: _type = "Task" [ 930.679675] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.689819] env[63021]: DEBUG oslo_vmware.api [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1293990, 'name': Rename_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.725171] env[63021]: DEBUG nova.network.neutron [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Updating instance_info_cache with network_info: [{"id": "43e48c53-d06d-41ff-97dd-a7747d5551d9", "address": "fa:16:3e:73:e1:9e", "network": {"id": "36548b96-bd4d-4725-8dab-5cb1cca12c4e", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-672805719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.178", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b4c8cd929ff4068b45ee829b19f9769", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8614db14-cc04-466b-b309-367ab8296cda", "external-id": "nsx-vlan-transportzone-337", "segmentation_id": 337, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43e48c53-d0", "ovs_interfaceid": "43e48c53-d06d-41ff-97dd-a7747d5551d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.772794] env[63021]: DEBUG nova.compute.manager [req-4ff874f5-4bcf-414b-87c5-e717c775fbd5 req-82d90de7-22cc-40ce-b466-c204cecd8318 service nova] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Received event network-vif-deleted-1019be13-e855-4eab-8201-a2ae2b22476c {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 930.773046] env[63021]: DEBUG nova.compute.manager [req-4ff874f5-4bcf-414b-87c5-e717c775fbd5 req-82d90de7-22cc-40ce-b466-c204cecd8318 service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Received event network-vif-plugged-eab7f4fe-4832-40e4-aed2-0e931d91a4d7 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 930.773203] env[63021]: DEBUG oslo_concurrency.lockutils [req-4ff874f5-4bcf-414b-87c5-e717c775fbd5 req-82d90de7-22cc-40ce-b466-c204cecd8318 service nova] Acquiring lock "2d647028-306a-41d5-96f0-2e82a98ab56d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.773422] env[63021]: DEBUG oslo_concurrency.lockutils [req-4ff874f5-4bcf-414b-87c5-e717c775fbd5 req-82d90de7-22cc-40ce-b466-c204cecd8318 service nova] Lock "2d647028-306a-41d5-96f0-2e82a98ab56d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.773582] env[63021]: DEBUG oslo_concurrency.lockutils [req-4ff874f5-4bcf-414b-87c5-e717c775fbd5 req-82d90de7-22cc-40ce-b466-c204cecd8318 service nova] Lock "2d647028-306a-41d5-96f0-2e82a98ab56d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.773747] env[63021]: DEBUG nova.compute.manager [req-4ff874f5-4bcf-414b-87c5-e717c775fbd5 req-82d90de7-22cc-40ce-b466-c204cecd8318 service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] No waiting events found dispatching network-vif-plugged-eab7f4fe-4832-40e4-aed2-0e931d91a4d7 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 930.773915] env[63021]: WARNING nova.compute.manager [req-4ff874f5-4bcf-414b-87c5-e717c775fbd5 req-82d90de7-22cc-40ce-b466-c204cecd8318 service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Received unexpected event network-vif-plugged-eab7f4fe-4832-40e4-aed2-0e931d91a4d7 for instance with vm_state building and task_state spawning. [ 930.774253] env[63021]: DEBUG nova.compute.manager [req-4ff874f5-4bcf-414b-87c5-e717c775fbd5 req-82d90de7-22cc-40ce-b466-c204cecd8318 service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Received event network-changed-eab7f4fe-4832-40e4-aed2-0e931d91a4d7 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 930.774448] env[63021]: DEBUG nova.compute.manager [req-4ff874f5-4bcf-414b-87c5-e717c775fbd5 req-82d90de7-22cc-40ce-b466-c204cecd8318 service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Refreshing instance network info cache due to event network-changed-eab7f4fe-4832-40e4-aed2-0e931d91a4d7. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 930.774627] env[63021]: DEBUG oslo_concurrency.lockutils [req-4ff874f5-4bcf-414b-87c5-e717c775fbd5 req-82d90de7-22cc-40ce-b466-c204cecd8318 service nova] Acquiring lock "refresh_cache-2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.981022] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 930.981022] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Deleting contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 930.981022] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Deleting the datastore file [datastore2] df0b6974-7f78-4b73-8583-d09754fbacea {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 930.981022] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0f98c51a-3935-48b5-9e18-ed91348b12bf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.986237] env[63021]: DEBUG oslo_vmware.api [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for the task: (returnval){ [ 930.986237] env[63021]: value = "task-1293991" [ 930.986237] env[63021]: _type = "Task" [ 930.986237] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.995138] env[63021]: DEBUG oslo_vmware.api [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293991, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.013265] env[63021]: DEBUG nova.scheduler.client.report [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 931.046099] env[63021]: DEBUG oslo_vmware.api [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1293983, 'name': PowerOnVM_Task, 'duration_secs': 1.032793} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.046443] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 931.046694] env[63021]: INFO nova.compute.manager [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Took 9.42 seconds to spawn the instance on the hypervisor. [ 931.046919] env[63021]: DEBUG nova.compute.manager [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 931.047729] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a55cb04-9dd8-48c3-8bd2-b4a0759081e5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.112131] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d0020b0d-18e8-4912-9dc0-605a5e814d35 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.175075] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Acquiring lock "a52e4cb5-d0ea-4698-9955-753626762a02" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.175396] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Lock "a52e4cb5-d0ea-4698-9955-753626762a02" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.185190] env[63021]: DEBUG nova.network.neutron [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 931.192118] env[63021]: DEBUG oslo_vmware.api [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1293990, 'name': Rename_Task, 'duration_secs': 0.354034} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.192708] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 931.193188] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-06e9d310-02ed-41eb-8025-3774bfad09ef {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.200579] env[63021]: DEBUG oslo_vmware.api [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 931.200579] env[63021]: value = "task-1293992" [ 931.200579] env[63021]: _type = "Task" [ 931.200579] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.208466] env[63021]: DEBUG oslo_vmware.api [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1293992, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.227058] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Releasing lock "refresh_cache-45c9e2be-eb19-4b83-b280-c9eeaddfccba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.414784] env[63021]: DEBUG nova.network.neutron [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Updating instance_info_cache with network_info: [{"id": "eab7f4fe-4832-40e4-aed2-0e931d91a4d7", "address": "fa:16:3e:9c:24:88", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeab7f4fe-48", "ovs_interfaceid": "eab7f4fe-4832-40e4-aed2-0e931d91a4d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.496757] env[63021]: DEBUG oslo_vmware.api [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Task: {'id': task-1293991, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.473337} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.497048] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 931.497238] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Deleted contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 931.497419] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 931.497593] env[63021]: INFO nova.compute.manager [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Took 1.46 seconds to destroy the instance on the hypervisor. [ 931.497913] env[63021]: DEBUG oslo.service.loopingcall [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 931.498121] env[63021]: DEBUG nova.compute.manager [-] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 931.498214] env[63021]: DEBUG nova.network.neutron [-] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 931.517911] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.814s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.520331] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.634s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.522202] env[63021]: INFO nova.compute.claims [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 931.566884] env[63021]: INFO nova.compute.manager [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Took 22.74 seconds to build instance. [ 931.579402] env[63021]: INFO nova.scheduler.client.report [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Deleted allocations for instance 4d712cc1-6202-407e-8767-c620a1ce0bba [ 931.677803] env[63021]: DEBUG nova.compute.manager [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 931.710816] env[63021]: DEBUG oslo_vmware.api [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1293992, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.775300] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 931.775300] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d74d13d0-8b97-4eee-ab78-558570686711 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.788231] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 931.788231] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-45b92230-a254-4303-8527-47a8f741d625 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.853091] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 931.853091] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 931.853091] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Deleting the datastore file [datastore1] 45c9e2be-eb19-4b83-b280-c9eeaddfccba {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 931.853091] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e9c9f9ce-a3ee-4261-b28a-78bb19e716d3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.860607] env[63021]: DEBUG oslo_vmware.api [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Waiting for the task: (returnval){ [ 931.860607] env[63021]: value = "task-1293995" [ 931.860607] env[63021]: _type = "Task" [ 931.860607] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.869678] env[63021]: DEBUG oslo_vmware.api [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1293995, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.917937] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Releasing lock "refresh_cache-2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.918427] env[63021]: DEBUG nova.compute.manager [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Instance network_info: |[{"id": "eab7f4fe-4832-40e4-aed2-0e931d91a4d7", "address": "fa:16:3e:9c:24:88", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeab7f4fe-48", "ovs_interfaceid": "eab7f4fe-4832-40e4-aed2-0e931d91a4d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 931.918910] env[63021]: DEBUG oslo_concurrency.lockutils [req-4ff874f5-4bcf-414b-87c5-e717c775fbd5 req-82d90de7-22cc-40ce-b466-c204cecd8318 service nova] Acquired lock "refresh_cache-2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.919162] env[63021]: DEBUG nova.network.neutron [req-4ff874f5-4bcf-414b-87c5-e717c775fbd5 req-82d90de7-22cc-40ce-b466-c204cecd8318 service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Refreshing network info cache for port eab7f4fe-4832-40e4-aed2-0e931d91a4d7 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 931.920487] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9c:24:88', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8059554c-499f-44b4-be06-29f80ec36b34', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eab7f4fe-4832-40e4-aed2-0e931d91a4d7', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 931.930779] env[63021]: DEBUG oslo.service.loopingcall [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 931.931664] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 931.932053] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9df2a68d-515a-425d-8537-91a2521e2067 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.959149] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 931.959149] env[63021]: value = "task-1293996" [ 931.959149] env[63021]: _type = "Task" [ 931.959149] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.971464] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293996, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.068904] env[63021]: DEBUG oslo_concurrency.lockutils [None req-8fae3893-735a-4caf-b8ae-c92dd2b84899 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "ae103118-bddf-46fe-90b9-98b60952ebba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.250s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.088438] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b894783d-7851-4ee7-94da-ca3571ddb49e tempest-ServerShowV254Test-1571660612 tempest-ServerShowV254Test-1571660612-project-member] Lock "4d712cc1-6202-407e-8767-c620a1ce0bba" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.893s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.206962] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.212779] env[63021]: DEBUG oslo_vmware.api [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1293992, 'name': PowerOnVM_Task, 'duration_secs': 0.794441} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.213093] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 932.213299] env[63021]: INFO nova.compute.manager [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Took 8.20 seconds to spawn the instance on the hypervisor. [ 932.213510] env[63021]: DEBUG nova.compute.manager [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 932.214302] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94a2f49d-5d47-4fc7-bca7-9892b8420595 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.366439] env[63021]: DEBUG nova.network.neutron [-] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.371115] env[63021]: DEBUG oslo_vmware.api [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1293995, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145143} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.371606] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 932.372338] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 932.372600] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 932.397917] env[63021]: INFO nova.scheduler.client.report [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Deleted allocations for instance 45c9e2be-eb19-4b83-b280-c9eeaddfccba [ 932.469230] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1293996, 'name': CreateVM_Task, 'duration_secs': 0.479541} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.469414] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 932.470177] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.470379] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.470728] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 932.471526] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96f819da-f3eb-49fb-ba2b-f54132d02e24 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.480133] env[63021]: DEBUG oslo_vmware.api [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 932.480133] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]526de521-723b-b373-4e88-45512b674d21" [ 932.480133] env[63021]: _type = "Task" [ 932.480133] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.489143] env[63021]: DEBUG oslo_vmware.api [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]526de521-723b-b373-4e88-45512b674d21, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.679155] env[63021]: DEBUG nova.network.neutron [req-4ff874f5-4bcf-414b-87c5-e717c775fbd5 req-82d90de7-22cc-40ce-b466-c204cecd8318 service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Updated VIF entry in instance network info cache for port eab7f4fe-4832-40e4-aed2-0e931d91a4d7. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 932.679155] env[63021]: DEBUG nova.network.neutron [req-4ff874f5-4bcf-414b-87c5-e717c775fbd5 req-82d90de7-22cc-40ce-b466-c204cecd8318 service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Updating instance_info_cache with network_info: [{"id": "eab7f4fe-4832-40e4-aed2-0e931d91a4d7", "address": "fa:16:3e:9c:24:88", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeab7f4fe-48", "ovs_interfaceid": "eab7f4fe-4832-40e4-aed2-0e931d91a4d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.708151] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07cb0fc2-44f4-4015-a9db-5afe7261cd54 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.715601] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86a0fac3-94d6-4e72-a174-2351d1693809 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.754416] env[63021]: INFO nova.compute.manager [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Took 23.10 seconds to build instance. [ 932.755902] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e1db999-f276-4cf1-bb58-c6e69ef2351e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.763843] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81b5599c-d6b0-42c4-95b3-609a6387f7b8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.780113] env[63021]: DEBUG nova.compute.provider_tree [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 932.813463] env[63021]: DEBUG nova.compute.manager [req-57c8f525-a91c-496d-96fc-2deae2aedd62 req-d40ab8aa-dbef-4493-82ac-2e1b0f27dad9 service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Received event network-vif-unplugged-43e48c53-d06d-41ff-97dd-a7747d5551d9 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 932.813730] env[63021]: DEBUG oslo_concurrency.lockutils [req-57c8f525-a91c-496d-96fc-2deae2aedd62 req-d40ab8aa-dbef-4493-82ac-2e1b0f27dad9 service nova] Acquiring lock "45c9e2be-eb19-4b83-b280-c9eeaddfccba-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.813910] env[63021]: DEBUG oslo_concurrency.lockutils [req-57c8f525-a91c-496d-96fc-2deae2aedd62 req-d40ab8aa-dbef-4493-82ac-2e1b0f27dad9 service nova] Lock "45c9e2be-eb19-4b83-b280-c9eeaddfccba-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.814112] env[63021]: DEBUG oslo_concurrency.lockutils [req-57c8f525-a91c-496d-96fc-2deae2aedd62 req-d40ab8aa-dbef-4493-82ac-2e1b0f27dad9 service nova] Lock "45c9e2be-eb19-4b83-b280-c9eeaddfccba-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.814291] env[63021]: DEBUG nova.compute.manager [req-57c8f525-a91c-496d-96fc-2deae2aedd62 req-d40ab8aa-dbef-4493-82ac-2e1b0f27dad9 service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] No waiting events found dispatching network-vif-unplugged-43e48c53-d06d-41ff-97dd-a7747d5551d9 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 932.814462] env[63021]: WARNING nova.compute.manager [req-57c8f525-a91c-496d-96fc-2deae2aedd62 req-d40ab8aa-dbef-4493-82ac-2e1b0f27dad9 service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Received unexpected event network-vif-unplugged-43e48c53-d06d-41ff-97dd-a7747d5551d9 for instance with vm_state shelved_offloaded and task_state None. [ 932.814660] env[63021]: DEBUG nova.compute.manager [req-57c8f525-a91c-496d-96fc-2deae2aedd62 req-d40ab8aa-dbef-4493-82ac-2e1b0f27dad9 service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Received event network-changed-43e48c53-d06d-41ff-97dd-a7747d5551d9 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 932.814805] env[63021]: DEBUG nova.compute.manager [req-57c8f525-a91c-496d-96fc-2deae2aedd62 req-d40ab8aa-dbef-4493-82ac-2e1b0f27dad9 service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Refreshing instance network info cache due to event network-changed-43e48c53-d06d-41ff-97dd-a7747d5551d9. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 932.815091] env[63021]: DEBUG oslo_concurrency.lockutils [req-57c8f525-a91c-496d-96fc-2deae2aedd62 req-d40ab8aa-dbef-4493-82ac-2e1b0f27dad9 service nova] Acquiring lock "refresh_cache-45c9e2be-eb19-4b83-b280-c9eeaddfccba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.815248] env[63021]: DEBUG oslo_concurrency.lockutils [req-57c8f525-a91c-496d-96fc-2deae2aedd62 req-d40ab8aa-dbef-4493-82ac-2e1b0f27dad9 service nova] Acquired lock "refresh_cache-45c9e2be-eb19-4b83-b280-c9eeaddfccba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.815459] env[63021]: DEBUG nova.network.neutron [req-57c8f525-a91c-496d-96fc-2deae2aedd62 req-d40ab8aa-dbef-4493-82ac-2e1b0f27dad9 service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Refreshing network info cache for port 43e48c53-d06d-41ff-97dd-a7747d5551d9 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 932.872386] env[63021]: INFO nova.compute.manager [-] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Took 1.37 seconds to deallocate network for instance. [ 932.901870] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.985147] env[63021]: INFO nova.compute.manager [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Rescuing [ 932.985147] env[63021]: DEBUG oslo_concurrency.lockutils [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "refresh_cache-418a394b-dbb2-47a2-96a9-a1caa4c4bb51" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.985147] env[63021]: DEBUG oslo_concurrency.lockutils [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquired lock "refresh_cache-418a394b-dbb2-47a2-96a9-a1caa4c4bb51" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.985319] env[63021]: DEBUG nova.network.neutron [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 932.993590] env[63021]: DEBUG oslo_vmware.api [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]526de521-723b-b373-4e88-45512b674d21, 'name': SearchDatastore_Task, 'duration_secs': 0.024007} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.993590] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.994013] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 932.994867] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.994867] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.994867] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 932.995077] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8c4a71b8-212b-45e3-936e-8ea4e0dacff0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.006684] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 933.006684] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 933.006684] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2ee5040-c3c1-4c34-904a-54bbbb9e979b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.011635] env[63021]: DEBUG oslo_vmware.api [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 933.011635] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52530c45-4127-0593-b833-3a7b89ff026d" [ 933.011635] env[63021]: _type = "Task" [ 933.011635] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.019885] env[63021]: DEBUG oslo_vmware.api [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52530c45-4127-0593-b833-3a7b89ff026d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.182188] env[63021]: DEBUG oslo_concurrency.lockutils [req-4ff874f5-4bcf-414b-87c5-e717c775fbd5 req-82d90de7-22cc-40ce-b466-c204cecd8318 service nova] Releasing lock "refresh_cache-2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.255273] env[63021]: DEBUG oslo_concurrency.lockutils [None req-81b0f7b2-8dea-415f-98e9-79154f31906f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "418a394b-dbb2-47a2-96a9-a1caa4c4bb51" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.070s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.283092] env[63021]: DEBUG nova.scheduler.client.report [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 933.378316] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.523652] env[63021]: DEBUG oslo_vmware.api [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52530c45-4127-0593-b833-3a7b89ff026d, 'name': SearchDatastore_Task, 'duration_secs': 0.009123} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.527036] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6ffb680-0c75-4b4e-bd62-bd6b4b047eb2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.534845] env[63021]: DEBUG oslo_vmware.api [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 933.534845] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5261deba-9955-94d7-eaaf-6130f95fbbb4" [ 933.534845] env[63021]: _type = "Task" [ 933.534845] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.543344] env[63021]: DEBUG oslo_vmware.api [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5261deba-9955-94d7-eaaf-6130f95fbbb4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.584707] env[63021]: DEBUG nova.network.neutron [req-57c8f525-a91c-496d-96fc-2deae2aedd62 req-d40ab8aa-dbef-4493-82ac-2e1b0f27dad9 service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Updated VIF entry in instance network info cache for port 43e48c53-d06d-41ff-97dd-a7747d5551d9. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 933.585316] env[63021]: DEBUG nova.network.neutron [req-57c8f525-a91c-496d-96fc-2deae2aedd62 req-d40ab8aa-dbef-4493-82ac-2e1b0f27dad9 service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Updating instance_info_cache with network_info: [{"id": "43e48c53-d06d-41ff-97dd-a7747d5551d9", "address": "fa:16:3e:73:e1:9e", "network": {"id": "36548b96-bd4d-4725-8dab-5cb1cca12c4e", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-672805719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.178", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b4c8cd929ff4068b45ee829b19f9769", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap43e48c53-d0", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.724830] env[63021]: DEBUG nova.network.neutron [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Updating instance_info_cache with network_info: [{"id": "905466dc-1047-469f-b282-768293b6d154", "address": "fa:16:3e:db:7b:17", "network": {"id": "3aaeed35-f452-4144-a259-8f440f8a8bf4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-229941363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b672b1d7d7ad454ebfa3e2fd2421c754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap905466dc-10", "ovs_interfaceid": "905466dc-1047-469f-b282-768293b6d154", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.787954] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.267s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.788330] env[63021]: DEBUG nova.compute.manager [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 933.790946] env[63021]: DEBUG oslo_concurrency.lockutils [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.562s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.791186] env[63021]: DEBUG nova.objects.instance [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lazy-loading 'resources' on Instance uuid d63fd465-975f-42ce-b1be-ac8a6929bc4d {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 933.917962] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-24e72953-af49-46ed-99ac-23f23801a0d0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Volume attach. Driver type: vmdk {{(pid=63021) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 933.918357] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-24e72953-af49-46ed-99ac-23f23801a0d0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277555', 'volume_id': '460ebecc-27ec-4047-99f6-a80a6e32f71d', 'name': 'volume-460ebecc-27ec-4047-99f6-a80a6e32f71d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '140184cd-6e4c-425c-8b17-361a1f565e93', 'attached_at': '', 'detached_at': '', 'volume_id': '460ebecc-27ec-4047-99f6-a80a6e32f71d', 'serial': '460ebecc-27ec-4047-99f6-a80a6e32f71d'} {{(pid=63021) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 933.919595] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2b8b58e-09c0-4cd6-92e4-0e4bc40b5588 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.942859] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-786586b9-b348-4849-8dc6-176577cb145a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.968043] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-24e72953-af49-46ed-99ac-23f23801a0d0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] volume-460ebecc-27ec-4047-99f6-a80a6e32f71d/volume-460ebecc-27ec-4047-99f6-a80a6e32f71d.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 933.968043] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-36ae0fbc-2f62-497e-9c6b-08e12f662a1c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.985544] env[63021]: DEBUG oslo_vmware.api [None req-24e72953-af49-46ed-99ac-23f23801a0d0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 933.985544] env[63021]: value = "task-1293997" [ 933.985544] env[63021]: _type = "Task" [ 933.985544] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.993566] env[63021]: DEBUG oslo_vmware.api [None req-24e72953-af49-46ed-99ac-23f23801a0d0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293997, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.046350] env[63021]: DEBUG oslo_vmware.api [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5261deba-9955-94d7-eaaf-6130f95fbbb4, 'name': SearchDatastore_Task, 'duration_secs': 0.00983} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.046656] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.046936] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 2d647028-306a-41d5-96f0-2e82a98ab56d/2d647028-306a-41d5-96f0-2e82a98ab56d.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 934.047352] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f1515790-48d5-4b2a-b7fe-f57b8baee431 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.054383] env[63021]: DEBUG oslo_vmware.api [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 934.054383] env[63021]: value = "task-1293998" [ 934.054383] env[63021]: _type = "Task" [ 934.054383] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.062718] env[63021]: DEBUG oslo_vmware.api [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293998, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.087954] env[63021]: DEBUG oslo_concurrency.lockutils [req-57c8f525-a91c-496d-96fc-2deae2aedd62 req-d40ab8aa-dbef-4493-82ac-2e1b0f27dad9 service nova] Releasing lock "refresh_cache-45c9e2be-eb19-4b83-b280-c9eeaddfccba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.088411] env[63021]: DEBUG nova.compute.manager [req-57c8f525-a91c-496d-96fc-2deae2aedd62 req-d40ab8aa-dbef-4493-82ac-2e1b0f27dad9 service nova] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Received event network-vif-deleted-948a6d57-d04a-4c5c-ae6f-cd5786ec9a89 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 934.228365] env[63021]: DEBUG oslo_concurrency.lockutils [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Releasing lock "refresh_cache-418a394b-dbb2-47a2-96a9-a1caa4c4bb51" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.294901] env[63021]: DEBUG nova.compute.utils [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 934.300520] env[63021]: DEBUG nova.compute.manager [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 934.300520] env[63021]: DEBUG nova.network.neutron [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 934.357361] env[63021]: DEBUG nova.policy [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2642402689b04f89946802ac3b3e6287', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '805af233efa147b3982fb794cd9e6f73', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 934.500101] env[63021]: DEBUG oslo_vmware.api [None req-24e72953-af49-46ed-99ac-23f23801a0d0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293997, 'name': ReconfigVM_Task, 'duration_secs': 0.497154} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.504938] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-24e72953-af49-46ed-99ac-23f23801a0d0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Reconfigured VM instance instance-00000051 to attach disk [datastore2] volume-460ebecc-27ec-4047-99f6-a80a6e32f71d/volume-460ebecc-27ec-4047-99f6-a80a6e32f71d.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 934.515863] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-585d595e-b1f2-4d73-b903-a9837677f8f9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.549535] env[63021]: DEBUG oslo_vmware.api [None req-24e72953-af49-46ed-99ac-23f23801a0d0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 934.549535] env[63021]: value = "task-1293999" [ 934.549535] env[63021]: _type = "Task" [ 934.549535] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.573116] env[63021]: DEBUG oslo_vmware.api [None req-24e72953-af49-46ed-99ac-23f23801a0d0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293999, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.582590] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8cc6167-cdb6-4ba9-8429-a3464900367f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.588798] env[63021]: DEBUG oslo_vmware.api [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293998, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.599449] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0788dd8b-90ab-4c5a-8ae8-43686dff8b4a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.640487] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16f22d96-1bd5-4a6e-9c6c-c3a3f76cdac0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.651119] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea9ee499-c892-4c19-a7b3-8ca63f15323d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.665057] env[63021]: DEBUG nova.compute.provider_tree [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 934.690378] env[63021]: DEBUG nova.network.neutron [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Successfully created port: a7ecacae-44cb-4b30-892f-d0f5ef3ad9d5 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 934.758295] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 934.758696] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4ba80265-a8d0-4fd8-b132-819d7c7279c7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.767261] env[63021]: DEBUG oslo_vmware.api [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 934.767261] env[63021]: value = "task-1294000" [ 934.767261] env[63021]: _type = "Task" [ 934.767261] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.779639] env[63021]: DEBUG oslo_vmware.api [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294000, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.806021] env[63021]: DEBUG nova.compute.manager [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 935.061661] env[63021]: DEBUG oslo_vmware.api [None req-24e72953-af49-46ed-99ac-23f23801a0d0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1293999, 'name': ReconfigVM_Task, 'duration_secs': 0.167361} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.065161] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-24e72953-af49-46ed-99ac-23f23801a0d0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277555', 'volume_id': '460ebecc-27ec-4047-99f6-a80a6e32f71d', 'name': 'volume-460ebecc-27ec-4047-99f6-a80a6e32f71d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '140184cd-6e4c-425c-8b17-361a1f565e93', 'attached_at': '', 'detached_at': '', 'volume_id': '460ebecc-27ec-4047-99f6-a80a6e32f71d', 'serial': '460ebecc-27ec-4047-99f6-a80a6e32f71d'} {{(pid=63021) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 935.071673] env[63021]: DEBUG oslo_vmware.api [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1293998, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.578206} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.071991] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 2d647028-306a-41d5-96f0-2e82a98ab56d/2d647028-306a-41d5-96f0-2e82a98ab56d.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 935.072261] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 935.073027] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7f4769da-c732-42de-a97a-4fab5a701050 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.082323] env[63021]: DEBUG oslo_vmware.api [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 935.082323] env[63021]: value = "task-1294001" [ 935.082323] env[63021]: _type = "Task" [ 935.082323] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.090144] env[63021]: DEBUG oslo_vmware.api [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294001, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.171077] env[63021]: DEBUG nova.scheduler.client.report [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 935.277314] env[63021]: DEBUG oslo_vmware.api [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294000, 'name': PowerOffVM_Task, 'duration_secs': 0.478877} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.277557] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 935.278915] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60ee653e-ec29-472e-bacd-1250ab0a165e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.297428] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff41b7c6-8cab-48b9-86a6-79976985a74f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.331471] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 935.331471] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6e4708b3-e1e3-46cd-b86f-d40c6db08ac4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.339150] env[63021]: DEBUG oslo_vmware.api [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 935.339150] env[63021]: value = "task-1294002" [ 935.339150] env[63021]: _type = "Task" [ 935.339150] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.339576] env[63021]: DEBUG oslo_concurrency.lockutils [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Acquiring lock "45c9e2be-eb19-4b83-b280-c9eeaddfccba" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.348680] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] VM already powered off {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 935.348923] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 935.349187] env[63021]: DEBUG oslo_concurrency.lockutils [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.349338] env[63021]: DEBUG oslo_concurrency.lockutils [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.349520] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 935.349779] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9c6412d3-6bb1-4493-a864-925305d6ea16 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.365683] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 935.365683] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 935.366230] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ad9803e-29c8-4391-a171-9a1d2c0b84b9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.371511] env[63021]: DEBUG oslo_vmware.api [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 935.371511] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]528d9661-385c-6b6a-277f-ebdb6a0ebe82" [ 935.371511] env[63021]: _type = "Task" [ 935.371511] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.379282] env[63021]: DEBUG oslo_vmware.api [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]528d9661-385c-6b6a-277f-ebdb6a0ebe82, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.591478] env[63021]: DEBUG oslo_vmware.api [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294001, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.676568] env[63021]: DEBUG oslo_concurrency.lockutils [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.885s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.679025] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d0020b0d-18e8-4912-9dc0-605a5e814d35 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.567s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.679164] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d0020b0d-18e8-4912-9dc0-605a5e814d35 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.681262] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.474s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.682731] env[63021]: INFO nova.compute.claims [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 935.703299] env[63021]: INFO nova.scheduler.client.report [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Deleted allocations for instance d63fd465-975f-42ce-b1be-ac8a6929bc4d [ 935.707663] env[63021]: INFO nova.scheduler.client.report [None req-d0020b0d-18e8-4912-9dc0-605a5e814d35 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Deleted allocations for instance 71909213-258d-45f5-9a3e-e473e8ce8aa2 [ 935.814392] env[63021]: DEBUG nova.compute.manager [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 935.836650] env[63021]: DEBUG nova.virt.hardware [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 935.836917] env[63021]: DEBUG nova.virt.hardware [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 935.837103] env[63021]: DEBUG nova.virt.hardware [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 935.837295] env[63021]: DEBUG nova.virt.hardware [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 935.837443] env[63021]: DEBUG nova.virt.hardware [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 935.837593] env[63021]: DEBUG nova.virt.hardware [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 935.837817] env[63021]: DEBUG nova.virt.hardware [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 935.837997] env[63021]: DEBUG nova.virt.hardware [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 935.838200] env[63021]: DEBUG nova.virt.hardware [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 935.838354] env[63021]: DEBUG nova.virt.hardware [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 935.838528] env[63021]: DEBUG nova.virt.hardware [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 935.839551] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d674afa1-d392-4117-be7f-e740fc9e05d5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.847131] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afcec924-6c51-4d89-be12-f39136040ab6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.881264] env[63021]: DEBUG oslo_vmware.api [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]528d9661-385c-6b6a-277f-ebdb6a0ebe82, 'name': SearchDatastore_Task, 'duration_secs': 0.04803} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.882067] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-15001d14-f4d6-4d9a-90dc-e62f528007d1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.886917] env[63021]: DEBUG oslo_vmware.api [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 935.886917] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5254f985-a77f-383e-a512-3a586631b94d" [ 935.886917] env[63021]: _type = "Task" [ 935.886917] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.894255] env[63021]: DEBUG oslo_vmware.api [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5254f985-a77f-383e-a512-3a586631b94d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.094707] env[63021]: DEBUG oslo_vmware.api [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294001, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.667952} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.094707] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 936.095026] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-862a8f93-1ab8-4c94-b0da-f500efa08023 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.109626] env[63021]: DEBUG nova.objects.instance [None req-24e72953-af49-46ed-99ac-23f23801a0d0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lazy-loading 'flavor' on Instance uuid 140184cd-6e4c-425c-8b17-361a1f565e93 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 936.124613] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] 2d647028-306a-41d5-96f0-2e82a98ab56d/2d647028-306a-41d5-96f0-2e82a98ab56d.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 936.126649] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c7cf3502-9676-4d6c-92e6-aef68d61b3b1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.142349] env[63021]: DEBUG nova.compute.manager [req-58a72ab0-115a-41aa-8fd7-a5503275ad0d req-61fc3d7e-1d78-410f-baf6-32750a7e6609 service nova] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Received event network-vif-plugged-a7ecacae-44cb-4b30-892f-d0f5ef3ad9d5 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 936.142573] env[63021]: DEBUG oslo_concurrency.lockutils [req-58a72ab0-115a-41aa-8fd7-a5503275ad0d req-61fc3d7e-1d78-410f-baf6-32750a7e6609 service nova] Acquiring lock "207c8197-6191-4837-8ead-8c24aa5b35bc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.142787] env[63021]: DEBUG oslo_concurrency.lockutils [req-58a72ab0-115a-41aa-8fd7-a5503275ad0d req-61fc3d7e-1d78-410f-baf6-32750a7e6609 service nova] Lock "207c8197-6191-4837-8ead-8c24aa5b35bc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.142956] env[63021]: DEBUG oslo_concurrency.lockutils [req-58a72ab0-115a-41aa-8fd7-a5503275ad0d req-61fc3d7e-1d78-410f-baf6-32750a7e6609 service nova] Lock "207c8197-6191-4837-8ead-8c24aa5b35bc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.143148] env[63021]: DEBUG nova.compute.manager [req-58a72ab0-115a-41aa-8fd7-a5503275ad0d req-61fc3d7e-1d78-410f-baf6-32750a7e6609 service nova] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] No waiting events found dispatching network-vif-plugged-a7ecacae-44cb-4b30-892f-d0f5ef3ad9d5 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 936.143310] env[63021]: WARNING nova.compute.manager [req-58a72ab0-115a-41aa-8fd7-a5503275ad0d req-61fc3d7e-1d78-410f-baf6-32750a7e6609 service nova] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Received unexpected event network-vif-plugged-a7ecacae-44cb-4b30-892f-d0f5ef3ad9d5 for instance with vm_state building and task_state spawning. [ 936.149656] env[63021]: DEBUG oslo_vmware.api [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 936.149656] env[63021]: value = "task-1294003" [ 936.149656] env[63021]: _type = "Task" [ 936.149656] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.157946] env[63021]: DEBUG oslo_vmware.api [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294003, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.207573] env[63021]: DEBUG nova.network.neutron [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Successfully updated port: a7ecacae-44cb-4b30-892f-d0f5ef3ad9d5 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 936.213774] env[63021]: DEBUG oslo_concurrency.lockutils [None req-436486a3-6d6f-4b97-b587-fa4797f7bdb2 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "d63fd465-975f-42ce-b1be-ac8a6929bc4d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.278s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.214664] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d0020b0d-18e8-4912-9dc0-605a5e814d35 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "71909213-258d-45f5-9a3e-e473e8ce8aa2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.063s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.399668] env[63021]: DEBUG oslo_vmware.api [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5254f985-a77f-383e-a512-3a586631b94d, 'name': SearchDatastore_Task, 'duration_secs': 0.009405} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.399973] env[63021]: DEBUG oslo_concurrency.lockutils [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.400298] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 418a394b-dbb2-47a2-96a9-a1caa4c4bb51/e624225d-af0f-415c-90cc-d4b228c479f6-rescue.vmdk. {{(pid=63021) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 936.400599] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a017210f-33f9-4c77-a3f5-e55bd5fa41dc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.406849] env[63021]: DEBUG oslo_vmware.api [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 936.406849] env[63021]: value = "task-1294004" [ 936.406849] env[63021]: _type = "Task" [ 936.406849] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.414374] env[63021]: DEBUG oslo_vmware.api [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294004, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.629437] env[63021]: DEBUG oslo_concurrency.lockutils [None req-24e72953-af49-46ed-99ac-23f23801a0d0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "140184cd-6e4c-425c-8b17-361a1f565e93" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.337s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.661630] env[63021]: DEBUG oslo_vmware.api [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294003, 'name': ReconfigVM_Task, 'duration_secs': 0.505634} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.661963] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Reconfigured VM instance instance-00000054 to attach disk [datastore1] 2d647028-306a-41d5-96f0-2e82a98ab56d/2d647028-306a-41d5-96f0-2e82a98ab56d.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 936.662723] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b0fd39a8-17f5-40fb-a0eb-62404d375fd2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.671923] env[63021]: DEBUG oslo_vmware.api [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 936.671923] env[63021]: value = "task-1294005" [ 936.671923] env[63021]: _type = "Task" [ 936.671923] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.680554] env[63021]: DEBUG oslo_vmware.api [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294005, 'name': Rename_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.712742] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "refresh_cache-207c8197-6191-4837-8ead-8c24aa5b35bc" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.712742] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquired lock "refresh_cache-207c8197-6191-4837-8ead-8c24aa5b35bc" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.712886] env[63021]: DEBUG nova.network.neutron [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 936.724662] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "140184cd-6e4c-425c-8b17-361a1f565e93" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.724949] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "140184cd-6e4c-425c-8b17-361a1f565e93" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.725171] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "140184cd-6e4c-425c-8b17-361a1f565e93-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.725354] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "140184cd-6e4c-425c-8b17-361a1f565e93-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.725614] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "140184cd-6e4c-425c-8b17-361a1f565e93-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.727836] env[63021]: INFO nova.compute.manager [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Terminating instance [ 936.733086] env[63021]: DEBUG nova.compute.manager [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 936.733271] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 936.733598] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-409f6a1d-a65b-452d-acb9-dd5d7ce42779 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.742058] env[63021]: DEBUG oslo_vmware.api [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 936.742058] env[63021]: value = "task-1294006" [ 936.742058] env[63021]: _type = "Task" [ 936.742058] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.754239] env[63021]: DEBUG oslo_vmware.api [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294006, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.892112] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9107e557-5fc5-46b5-9c97-becf131aa8e9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.900771] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd336368-74b5-44dc-96a9-8dda5ab50b8b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.937355] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c79db690-5706-41cd-b02d-bf8eb446b330 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.945586] env[63021]: DEBUG oslo_vmware.api [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294004, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.497209} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.946328] env[63021]: INFO nova.virt.vmwareapi.ds_util [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 418a394b-dbb2-47a2-96a9-a1caa4c4bb51/e624225d-af0f-415c-90cc-d4b228c479f6-rescue.vmdk. [ 936.947365] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46a7ec43-f645-4f50-b386-753b3c3d1214 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.956575] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6431b2d0-9f8f-45f3-b7e2-ccf3a0425ddf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.980436] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] 418a394b-dbb2-47a2-96a9-a1caa4c4bb51/e624225d-af0f-415c-90cc-d4b228c479f6-rescue.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 936.984028] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5cec6719-e1b7-4427-8ff4-905ff43c5750 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.003219] env[63021]: DEBUG nova.compute.provider_tree [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 937.011656] env[63021]: DEBUG oslo_vmware.api [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 937.011656] env[63021]: value = "task-1294007" [ 937.011656] env[63021]: _type = "Task" [ 937.011656] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.021541] env[63021]: DEBUG oslo_vmware.api [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294007, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.138029] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "881c2fb8-5f8c-48c2-a173-8fab949f24f7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.138442] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "881c2fb8-5f8c-48c2-a173-8fab949f24f7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.138753] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "881c2fb8-5f8c-48c2-a173-8fab949f24f7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.139074] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "881c2fb8-5f8c-48c2-a173-8fab949f24f7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.139328] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "881c2fb8-5f8c-48c2-a173-8fab949f24f7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.141665] env[63021]: INFO nova.compute.manager [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Terminating instance [ 937.143641] env[63021]: DEBUG nova.compute.manager [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 937.143899] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 937.144752] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68b84285-deb7-4c91-b033-ac62505d432a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.152423] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 937.152778] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bafc39b4-f24d-42e1-8272-0f0c3bb42b3a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.158217] env[63021]: DEBUG oslo_vmware.api [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 937.158217] env[63021]: value = "task-1294008" [ 937.158217] env[63021]: _type = "Task" [ 937.158217] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.165736] env[63021]: DEBUG oslo_vmware.api [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294008, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.181296] env[63021]: DEBUG oslo_vmware.api [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294005, 'name': Rename_Task, 'duration_secs': 0.172478} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.181572] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 937.181817] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b83aa52d-e3ea-4ce5-aa89-abdfd923dcad {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.187526] env[63021]: DEBUG oslo_vmware.api [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 937.187526] env[63021]: value = "task-1294009" [ 937.187526] env[63021]: _type = "Task" [ 937.187526] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.195359] env[63021]: DEBUG oslo_vmware.api [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294009, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.253584] env[63021]: DEBUG oslo_vmware.api [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294006, 'name': PowerOffVM_Task, 'duration_secs': 0.207778} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.253584] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 937.253584] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Volume detach. Driver type: vmdk {{(pid=63021) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 937.253584] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277555', 'volume_id': '460ebecc-27ec-4047-99f6-a80a6e32f71d', 'name': 'volume-460ebecc-27ec-4047-99f6-a80a6e32f71d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '140184cd-6e4c-425c-8b17-361a1f565e93', 'attached_at': '', 'detached_at': '', 'volume_id': '460ebecc-27ec-4047-99f6-a80a6e32f71d', 'serial': '460ebecc-27ec-4047-99f6-a80a6e32f71d'} {{(pid=63021) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 937.254511] env[63021]: DEBUG nova.network.neutron [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 937.256983] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79b1b8e5-05f8-45e7-b91c-238fe6e9a63e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.284030] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea856761-7a7a-4846-b93e-886cde123a73 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.291178] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d076003-7f40-467f-b787-5ab8ccf0b8c4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.312618] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aed51831-8268-4fb0-89ce-91d096590572 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.330238] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] The volume has not been displaced from its original location: [datastore2] volume-460ebecc-27ec-4047-99f6-a80a6e32f71d/volume-460ebecc-27ec-4047-99f6-a80a6e32f71d.vmdk. No consolidation needed. {{(pid=63021) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 937.335438] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Reconfiguring VM instance instance-00000051 to detach disk 2001 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 937.335766] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1ead2259-bcd0-4dc5-b7ed-bccc30da352a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.354852] env[63021]: DEBUG oslo_vmware.api [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 937.354852] env[63021]: value = "task-1294010" [ 937.354852] env[63021]: _type = "Task" [ 937.354852] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.362666] env[63021]: DEBUG oslo_vmware.api [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294010, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.428320] env[63021]: DEBUG nova.network.neutron [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Updating instance_info_cache with network_info: [{"id": "a7ecacae-44cb-4b30-892f-d0f5ef3ad9d5", "address": "fa:16:3e:b7:1c:44", "network": {"id": "708d0dc3-c651-4001-9638-1d3538ff76c9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-983471056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "805af233efa147b3982fb794cd9e6f73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7ecacae-44", "ovs_interfaceid": "a7ecacae-44cb-4b30-892f-d0f5ef3ad9d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.507790] env[63021]: DEBUG nova.scheduler.client.report [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 937.521209] env[63021]: DEBUG oslo_vmware.api [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294007, 'name': ReconfigVM_Task, 'duration_secs': 0.299864} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.521493] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Reconfigured VM instance instance-00000053 to attach disk [datastore1] 418a394b-dbb2-47a2-96a9-a1caa4c4bb51/e624225d-af0f-415c-90cc-d4b228c479f6-rescue.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 937.522342] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74e405b9-1ded-41fb-9740-b8f6d3895273 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.547318] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7ce3fc3b-ec99-4eba-883b-21216dca5498 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.562117] env[63021]: DEBUG oslo_vmware.api [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 937.562117] env[63021]: value = "task-1294011" [ 937.562117] env[63021]: _type = "Task" [ 937.562117] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.569791] env[63021]: DEBUG oslo_vmware.api [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294011, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.668143] env[63021]: DEBUG oslo_vmware.api [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294008, 'name': PowerOffVM_Task, 'duration_secs': 0.163153} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.669500] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 937.669683] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 937.672152] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f1866906-b7a4-488d-98df-9ae319812807 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.674078] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "919887c8-171d-48fe-8c6b-2abf5f55d6ab" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.674309] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "919887c8-171d-48fe-8c6b-2abf5f55d6ab" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.698590] env[63021]: DEBUG oslo_vmware.api [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294009, 'name': PowerOnVM_Task} progress is 81%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.820612] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 937.820993] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Deleting contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 937.821334] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Deleting the datastore file [datastore2] 881c2fb8-5f8c-48c2-a173-8fab949f24f7 {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 937.822193] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a86ca97c-2256-4430-afed-73e9a557f710 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.830141] env[63021]: DEBUG oslo_vmware.api [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 937.830141] env[63021]: value = "task-1294013" [ 937.830141] env[63021]: _type = "Task" [ 937.830141] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.841398] env[63021]: DEBUG oslo_vmware.api [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294013, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.864555] env[63021]: DEBUG oslo_vmware.api [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294010, 'name': ReconfigVM_Task, 'duration_secs': 0.320495} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.864802] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Reconfigured VM instance instance-00000051 to detach disk 2001 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 937.869457] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dc1207e3-387e-47c2-a64c-37202e8aac15 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.885622] env[63021]: DEBUG oslo_vmware.api [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 937.885622] env[63021]: value = "task-1294014" [ 937.885622] env[63021]: _type = "Task" [ 937.885622] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.893304] env[63021]: DEBUG oslo_vmware.api [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294014, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.931131] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Releasing lock "refresh_cache-207c8197-6191-4837-8ead-8c24aa5b35bc" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.931478] env[63021]: DEBUG nova.compute.manager [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Instance network_info: |[{"id": "a7ecacae-44cb-4b30-892f-d0f5ef3ad9d5", "address": "fa:16:3e:b7:1c:44", "network": {"id": "708d0dc3-c651-4001-9638-1d3538ff76c9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-983471056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "805af233efa147b3982fb794cd9e6f73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7ecacae-44", "ovs_interfaceid": "a7ecacae-44cb-4b30-892f-d0f5ef3ad9d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 937.931916] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b7:1c:44', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b7a73c01-1bb9-4612-a1a7-16d71b732e81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a7ecacae-44cb-4b30-892f-d0f5ef3ad9d5', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 937.939676] env[63021]: DEBUG oslo.service.loopingcall [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 937.940394] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 937.940632] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-12164594-99ae-42a1-a8f6-fe9fbb131201 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.959971] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 937.959971] env[63021]: value = "task-1294015" [ 937.959971] env[63021]: _type = "Task" [ 937.959971] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.967787] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294015, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.012960] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.332s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.013656] env[63021]: DEBUG nova.compute.manager [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 938.016422] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.115s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.016701] env[63021]: DEBUG nova.objects.instance [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lazy-loading 'resources' on Instance uuid 45c9e2be-eb19-4b83-b280-c9eeaddfccba {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 938.075071] env[63021]: DEBUG oslo_vmware.api [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294011, 'name': ReconfigVM_Task, 'duration_secs': 0.323102} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.075596] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 938.075914] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9bea4d84-ee1f-4ea9-abe8-d405acc05556 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.082803] env[63021]: DEBUG oslo_vmware.api [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 938.082803] env[63021]: value = "task-1294016" [ 938.082803] env[63021]: _type = "Task" [ 938.082803] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.094506] env[63021]: DEBUG oslo_vmware.api [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294016, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.177484] env[63021]: DEBUG nova.compute.manager [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 938.200671] env[63021]: DEBUG oslo_vmware.api [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294009, 'name': PowerOnVM_Task, 'duration_secs': 0.788957} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.200671] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 938.200671] env[63021]: INFO nova.compute.manager [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Took 9.38 seconds to spawn the instance on the hypervisor. [ 938.200671] env[63021]: DEBUG nova.compute.manager [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 938.200671] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92a0cb51-06c9-48e3-910c-04058dc3bb0f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.206962] env[63021]: DEBUG nova.compute.manager [req-3aeb09b7-77e2-4194-8448-bb4466c53746 req-da7dd7b3-ed7e-4e80-ac7a-ef60997926f9 service nova] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Received event network-changed-a7ecacae-44cb-4b30-892f-d0f5ef3ad9d5 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 938.206962] env[63021]: DEBUG nova.compute.manager [req-3aeb09b7-77e2-4194-8448-bb4466c53746 req-da7dd7b3-ed7e-4e80-ac7a-ef60997926f9 service nova] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Refreshing instance network info cache due to event network-changed-a7ecacae-44cb-4b30-892f-d0f5ef3ad9d5. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 938.206962] env[63021]: DEBUG oslo_concurrency.lockutils [req-3aeb09b7-77e2-4194-8448-bb4466c53746 req-da7dd7b3-ed7e-4e80-ac7a-ef60997926f9 service nova] Acquiring lock "refresh_cache-207c8197-6191-4837-8ead-8c24aa5b35bc" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.206962] env[63021]: DEBUG oslo_concurrency.lockutils [req-3aeb09b7-77e2-4194-8448-bb4466c53746 req-da7dd7b3-ed7e-4e80-ac7a-ef60997926f9 service nova] Acquired lock "refresh_cache-207c8197-6191-4837-8ead-8c24aa5b35bc" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.207530] env[63021]: DEBUG nova.network.neutron [req-3aeb09b7-77e2-4194-8448-bb4466c53746 req-da7dd7b3-ed7e-4e80-ac7a-ef60997926f9 service nova] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Refreshing network info cache for port a7ecacae-44cb-4b30-892f-d0f5ef3ad9d5 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 938.343047] env[63021]: DEBUG oslo_vmware.api [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294013, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.343432} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.343351] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 938.343540] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Deleted contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 938.343715] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 938.343885] env[63021]: INFO nova.compute.manager [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Took 1.20 seconds to destroy the instance on the hypervisor. [ 938.344135] env[63021]: DEBUG oslo.service.loopingcall [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 938.344850] env[63021]: DEBUG nova.compute.manager [-] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 938.344981] env[63021]: DEBUG nova.network.neutron [-] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 938.397060] env[63021]: DEBUG oslo_vmware.api [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294014, 'name': ReconfigVM_Task, 'duration_secs': 0.158462} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.397060] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277555', 'volume_id': '460ebecc-27ec-4047-99f6-a80a6e32f71d', 'name': 'volume-460ebecc-27ec-4047-99f6-a80a6e32f71d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '140184cd-6e4c-425c-8b17-361a1f565e93', 'attached_at': '', 'detached_at': '', 'volume_id': '460ebecc-27ec-4047-99f6-a80a6e32f71d', 'serial': '460ebecc-27ec-4047-99f6-a80a6e32f71d'} {{(pid=63021) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 938.397060] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 938.397637] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec151980-0b63-4439-911a-305d2ac4710d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.404307] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 938.404565] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a7bfd9b8-9513-4804-8cb1-e29590726359 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.470794] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294015, 'name': CreateVM_Task, 'duration_secs': 0.476942} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.471214] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 938.471927] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.472074] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.472389] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 938.472650] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-64a3cb8b-6048-4942-8d40-174f65ab08fa {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.479355] env[63021]: DEBUG oslo_vmware.api [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 938.479355] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]526c618b-98a5-8f17-15de-4adabac3c6f2" [ 938.479355] env[63021]: _type = "Task" [ 938.479355] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.484071] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 938.484290] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 938.484475] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Deleting the datastore file [datastore1] 140184cd-6e4c-425c-8b17-361a1f565e93 {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 938.484778] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0bc37f75-617f-478c-9cf0-00d61612260b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.491548] env[63021]: DEBUG oslo_vmware.api [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]526c618b-98a5-8f17-15de-4adabac3c6f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.496093] env[63021]: DEBUG oslo_vmware.api [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 938.496093] env[63021]: value = "task-1294018" [ 938.496093] env[63021]: _type = "Task" [ 938.496093] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.503448] env[63021]: DEBUG oslo_vmware.api [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294018, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.520010] env[63021]: DEBUG nova.compute.utils [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 938.521405] env[63021]: DEBUG nova.compute.manager [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 938.521547] env[63021]: DEBUG nova.network.neutron [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 938.523700] env[63021]: DEBUG nova.objects.instance [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lazy-loading 'numa_topology' on Instance uuid 45c9e2be-eb19-4b83-b280-c9eeaddfccba {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 938.576300] env[63021]: DEBUG nova.policy [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ec602745ec2847d6bcee3e9921317110', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f659d3d817cc4d7c8d718e5374acb4cb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 938.591890] env[63021]: DEBUG oslo_vmware.api [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294016, 'name': PowerOnVM_Task} progress is 76%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.710747] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.726687] env[63021]: INFO nova.compute.manager [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Took 23.08 seconds to build instance. [ 938.986927] env[63021]: DEBUG nova.network.neutron [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Successfully created port: 45d08a72-44cf-4326-b881-7b60e5d8d962 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 938.994739] env[63021]: DEBUG oslo_vmware.api [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]526c618b-98a5-8f17-15de-4adabac3c6f2, 'name': SearchDatastore_Task, 'duration_secs': 0.010089} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.995072] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.995316] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 938.995613] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.995775] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.996021] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 938.996338] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a777c3df-1ac7-4d1d-8b1a-85382023c5f0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.007145] env[63021]: DEBUG oslo_vmware.api [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294018, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162713} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.008343] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 939.008540] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 939.008722] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 939.008895] env[63021]: INFO nova.compute.manager [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Took 2.28 seconds to destroy the instance on the hypervisor. [ 939.009141] env[63021]: DEBUG oslo.service.loopingcall [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 939.009349] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 939.009505] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 939.010202] env[63021]: DEBUG nova.compute.manager [-] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 939.010302] env[63021]: DEBUG nova.network.neutron [-] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 939.012104] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7cbff2e-4058-4357-b6dd-967ecbf03acb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.017299] env[63021]: DEBUG oslo_vmware.api [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 939.017299] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52b935ee-34b2-850b-8f96-6025c9603c65" [ 939.017299] env[63021]: _type = "Task" [ 939.017299] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.025982] env[63021]: DEBUG nova.compute.manager [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 939.028426] env[63021]: DEBUG nova.objects.base [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Object Instance<45c9e2be-eb19-4b83-b280-c9eeaddfccba> lazy-loaded attributes: resources,numa_topology {{(pid=63021) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 939.031085] env[63021]: DEBUG oslo_vmware.api [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52b935ee-34b2-850b-8f96-6025c9603c65, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.048451] env[63021]: DEBUG nova.network.neutron [req-3aeb09b7-77e2-4194-8448-bb4466c53746 req-da7dd7b3-ed7e-4e80-ac7a-ef60997926f9 service nova] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Updated VIF entry in instance network info cache for port a7ecacae-44cb-4b30-892f-d0f5ef3ad9d5. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 939.048944] env[63021]: DEBUG nova.network.neutron [req-3aeb09b7-77e2-4194-8448-bb4466c53746 req-da7dd7b3-ed7e-4e80-ac7a-ef60997926f9 service nova] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Updating instance_info_cache with network_info: [{"id": "a7ecacae-44cb-4b30-892f-d0f5ef3ad9d5", "address": "fa:16:3e:b7:1c:44", "network": {"id": "708d0dc3-c651-4001-9638-1d3538ff76c9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-983471056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "805af233efa147b3982fb794cd9e6f73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7ecacae-44", "ovs_interfaceid": "a7ecacae-44cb-4b30-892f-d0f5ef3ad9d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.099714] env[63021]: DEBUG oslo_vmware.api [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294016, 'name': PowerOnVM_Task, 'duration_secs': 0.912734} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.100028] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 939.104919] env[63021]: DEBUG nova.compute.manager [None req-99b0d22b-e314-415d-99aa-221f11d3233d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 939.106138] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ce8a896-c0d0-4d47-9738-ac4799a7e5c4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.150275] env[63021]: DEBUG nova.network.neutron [-] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.229597] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f67fe548-3476-4fd2-b24a-cc86e72dd22d tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "2d647028-306a-41d5-96f0-2e82a98ab56d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.596s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.265020] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c79f39b-6f7c-4e2f-b706-19e6d489d3ec {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.274696] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec9615be-f9e8-4335-b26a-d55ce4bb31b5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.306497] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58ce370a-9bf9-4d4d-99b8-4c7f3473ac21 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.314914] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca9269e4-19fb-41d3-9825-c6817ac31ae1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.331223] env[63021]: DEBUG nova.compute.provider_tree [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 939.528305] env[63021]: DEBUG oslo_vmware.api [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52b935ee-34b2-850b-8f96-6025c9603c65, 'name': SearchDatastore_Task, 'duration_secs': 0.008159} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.529108] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6242952f-a01c-4ba6-89f6-3ef76e3b2a41 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.538199] env[63021]: DEBUG oslo_vmware.api [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 939.538199] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52bbd0e8-eb73-e49d-755d-818c6ddd81a4" [ 939.538199] env[63021]: _type = "Task" [ 939.538199] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.548833] env[63021]: DEBUG oslo_vmware.api [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52bbd0e8-eb73-e49d-755d-818c6ddd81a4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.551470] env[63021]: DEBUG oslo_concurrency.lockutils [req-3aeb09b7-77e2-4194-8448-bb4466c53746 req-da7dd7b3-ed7e-4e80-ac7a-ef60997926f9 service nova] Releasing lock "refresh_cache-207c8197-6191-4837-8ead-8c24aa5b35bc" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.653573] env[63021]: INFO nova.compute.manager [-] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Took 1.31 seconds to deallocate network for instance. [ 939.721299] env[63021]: DEBUG nova.compute.manager [req-7550cf99-cfea-41cf-9a4b-19bf568e8bb4 req-62285461-831c-43b2-8469-26c64398b628 service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Received event network-changed-eab7f4fe-4832-40e4-aed2-0e931d91a4d7 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 939.721519] env[63021]: DEBUG nova.compute.manager [req-7550cf99-cfea-41cf-9a4b-19bf568e8bb4 req-62285461-831c-43b2-8469-26c64398b628 service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Refreshing instance network info cache due to event network-changed-eab7f4fe-4832-40e4-aed2-0e931d91a4d7. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 939.721748] env[63021]: DEBUG oslo_concurrency.lockutils [req-7550cf99-cfea-41cf-9a4b-19bf568e8bb4 req-62285461-831c-43b2-8469-26c64398b628 service nova] Acquiring lock "refresh_cache-2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.722865] env[63021]: DEBUG oslo_concurrency.lockutils [req-7550cf99-cfea-41cf-9a4b-19bf568e8bb4 req-62285461-831c-43b2-8469-26c64398b628 service nova] Acquired lock "refresh_cache-2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.722865] env[63021]: DEBUG nova.network.neutron [req-7550cf99-cfea-41cf-9a4b-19bf568e8bb4 req-62285461-831c-43b2-8469-26c64398b628 service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Refreshing network info cache for port eab7f4fe-4832-40e4-aed2-0e931d91a4d7 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 939.837668] env[63021]: DEBUG nova.scheduler.client.report [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 939.855201] env[63021]: INFO nova.compute.manager [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Rebuilding instance [ 939.895600] env[63021]: DEBUG nova.compute.manager [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 939.896522] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45d44065-9b3a-4b15-a690-ffd4c7456abb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.005225] env[63021]: DEBUG nova.network.neutron [-] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.041452] env[63021]: DEBUG nova.compute.manager [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 940.059122] env[63021]: DEBUG oslo_vmware.api [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52bbd0e8-eb73-e49d-755d-818c6ddd81a4, 'name': SearchDatastore_Task, 'duration_secs': 0.009608} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.059847] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.060189] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 207c8197-6191-4837-8ead-8c24aa5b35bc/207c8197-6191-4837-8ead-8c24aa5b35bc.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 940.062112] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bb4328ed-a1b8-4eb2-ba2b-96a77d278653 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.068305] env[63021]: DEBUG oslo_vmware.api [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 940.068305] env[63021]: value = "task-1294019" [ 940.068305] env[63021]: _type = "Task" [ 940.068305] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.078964] env[63021]: DEBUG nova.virt.hardware [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 940.079449] env[63021]: DEBUG nova.virt.hardware [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 940.079753] env[63021]: DEBUG nova.virt.hardware [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 940.080262] env[63021]: DEBUG nova.virt.hardware [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 940.080462] env[63021]: DEBUG nova.virt.hardware [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 940.080727] env[63021]: DEBUG nova.virt.hardware [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 940.081090] env[63021]: DEBUG nova.virt.hardware [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 940.081275] env[63021]: DEBUG nova.virt.hardware [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 940.081446] env[63021]: DEBUG nova.virt.hardware [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 940.081621] env[63021]: DEBUG nova.virt.hardware [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 940.081797] env[63021]: DEBUG nova.virt.hardware [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 940.082587] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-064cc923-5cfc-45c4-ad8f-7fe81b5e920e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.093820] env[63021]: DEBUG oslo_vmware.api [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294019, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.098502] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f523604-3cd4-4148-8d31-74a872690324 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.173400] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.235496] env[63021]: DEBUG nova.compute.manager [req-ed891bce-378f-4c57-95c3-54f723f15878 req-80428ebc-e50d-48b5-ae16-684d5a42261c service nova] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Received event network-vif-deleted-3dc162c2-36e3-4bb0-b7a4-3116c11c145a {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 940.343259] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.327s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.345977] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.968s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.346262] env[63021]: DEBUG nova.objects.instance [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Lazy-loading 'resources' on Instance uuid df0b6974-7f78-4b73-8583-d09754fbacea {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 940.413234] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 940.413549] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-21be3729-cdcf-437e-b56a-e0e8f3ffa9de {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.422215] env[63021]: DEBUG oslo_vmware.api [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Waiting for the task: (returnval){ [ 940.422215] env[63021]: value = "task-1294020" [ 940.422215] env[63021]: _type = "Task" [ 940.422215] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.432425] env[63021]: DEBUG oslo_vmware.api [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Task: {'id': task-1294020, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.508371] env[63021]: INFO nova.compute.manager [-] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Took 1.50 seconds to deallocate network for instance. [ 940.576764] env[63021]: DEBUG nova.network.neutron [req-7550cf99-cfea-41cf-9a4b-19bf568e8bb4 req-62285461-831c-43b2-8469-26c64398b628 service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Updated VIF entry in instance network info cache for port eab7f4fe-4832-40e4-aed2-0e931d91a4d7. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 940.577177] env[63021]: DEBUG nova.network.neutron [req-7550cf99-cfea-41cf-9a4b-19bf568e8bb4 req-62285461-831c-43b2-8469-26c64398b628 service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Updating instance_info_cache with network_info: [{"id": "eab7f4fe-4832-40e4-aed2-0e931d91a4d7", "address": "fa:16:3e:9c:24:88", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeab7f4fe-48", "ovs_interfaceid": "eab7f4fe-4832-40e4-aed2-0e931d91a4d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.584485] env[63021]: DEBUG oslo_vmware.api [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294019, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.50487} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.585167] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 207c8197-6191-4837-8ead-8c24aa5b35bc/207c8197-6191-4837-8ead-8c24aa5b35bc.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 940.585447] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 940.585873] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-33c00bd5-c919-418b-ae30-25b5c23654b8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.594255] env[63021]: DEBUG oslo_vmware.api [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 940.594255] env[63021]: value = "task-1294021" [ 940.594255] env[63021]: _type = "Task" [ 940.594255] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.606811] env[63021]: DEBUG oslo_vmware.api [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294021, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.767543] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "6155c098-e4a6-47e6-b343-4a77ca90eb2e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.767720] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "6155c098-e4a6-47e6-b343-4a77ca90eb2e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.769561] env[63021]: DEBUG nova.network.neutron [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Successfully updated port: 45d08a72-44cf-4326-b881-7b60e5d8d962 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 940.856220] env[63021]: DEBUG oslo_concurrency.lockutils [None req-6870c5ed-6ad7-45c1-bb96-eedbf2d7fa94 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lock "45c9e2be-eb19-4b83-b280-c9eeaddfccba" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 34.278s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.857056] env[63021]: DEBUG oslo_concurrency.lockutils [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lock "45c9e2be-eb19-4b83-b280-c9eeaddfccba" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 5.517s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.857653] env[63021]: INFO nova.compute.manager [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Unshelving [ 940.934105] env[63021]: DEBUG oslo_vmware.api [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Task: {'id': task-1294020, 'name': PowerOffVM_Task, 'duration_secs': 0.269614} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.934681] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 940.935451] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 940.935717] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a8e002a3-7177-4251-8a5e-15883f70b01e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.944658] env[63021]: DEBUG oslo_vmware.api [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Waiting for the task: (returnval){ [ 940.944658] env[63021]: value = "task-1294022" [ 940.944658] env[63021]: _type = "Task" [ 940.944658] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.952555] env[63021]: DEBUG oslo_vmware.api [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Task: {'id': task-1294022, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.038671] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d22c11ad-4b29-4400-9836-ac9a849c5b1e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.047747] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42d01053-39aa-42de-ab42-b80dedd4a406 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.084022] env[63021]: DEBUG oslo_concurrency.lockutils [req-7550cf99-cfea-41cf-9a4b-19bf568e8bb4 req-62285461-831c-43b2-8469-26c64398b628 service nova] Releasing lock "refresh_cache-2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.085728] env[63021]: INFO nova.compute.manager [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Took 0.58 seconds to detach 1 volumes for instance. [ 941.088083] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4db5a496-8815-4b86-89d6-63836ff98143 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.102103] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4987d0f7-a93e-4481-aa65-0429d6992844 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.109963] env[63021]: DEBUG oslo_vmware.api [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294021, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068446} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.110114] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 941.110834] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbcffcb0-0c81-44df-999f-3cf23e0a4a05 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.122600] env[63021]: DEBUG nova.compute.provider_tree [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 941.144408] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] 207c8197-6191-4837-8ead-8c24aa5b35bc/207c8197-6191-4837-8ead-8c24aa5b35bc.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 941.146136] env[63021]: DEBUG nova.scheduler.client.report [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 941.149227] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-828984ab-03bf-4b75-a1d0-a8fdd5be1ad5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.165176] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.819s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.167512] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.457s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.169201] env[63021]: INFO nova.compute.claims [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 941.177783] env[63021]: DEBUG oslo_vmware.api [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 941.177783] env[63021]: value = "task-1294023" [ 941.177783] env[63021]: _type = "Task" [ 941.177783] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.187025] env[63021]: DEBUG oslo_vmware.api [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294023, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.192125] env[63021]: INFO nova.scheduler.client.report [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Deleted allocations for instance df0b6974-7f78-4b73-8583-d09754fbacea [ 941.271987] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Acquiring lock "refresh_cache-a52e4cb5-d0ea-4698-9955-753626762a02" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.272188] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Acquired lock "refresh_cache-a52e4cb5-d0ea-4698-9955-753626762a02" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.272380] env[63021]: DEBUG nova.network.neutron [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 941.273711] env[63021]: DEBUG nova.compute.manager [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 941.457833] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] VM already powered off {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 941.458104] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Volume detach. Driver type: vmdk {{(pid=63021) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 941.458311] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277493', 'volume_id': 'e424adc5-a8b0-4409-b3e0-f3cdbc8fb712', 'name': 'volume-e424adc5-a8b0-4409-b3e0-f3cdbc8fb712', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '80b857c4-d9dd-4483-970f-a0f5f4bcf173', 'attached_at': '', 'detached_at': '', 'volume_id': 'e424adc5-a8b0-4409-b3e0-f3cdbc8fb712', 'serial': 'e424adc5-a8b0-4409-b3e0-f3cdbc8fb712'} {{(pid=63021) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 941.462128] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c8f4659-8049-4f5f-8d46-25024cc96878 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.484575] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ad35b2f-3a15-421f-b970-67d1177aad03 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.493024] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-895e471e-0018-45f6-9bd6-1a76e9516f2b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.510911] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7803e67-9aee-471d-8bdc-b0b38fa8e540 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.527283] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] The volume has not been displaced from its original location: [datastore2] volume-e424adc5-a8b0-4409-b3e0-f3cdbc8fb712/volume-e424adc5-a8b0-4409-b3e0-f3cdbc8fb712.vmdk. No consolidation needed. {{(pid=63021) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 941.532862] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Reconfiguring VM instance instance-0000004d to detach disk 2000 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 941.533244] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-224965b3-522d-4736-bbd2-f8c814765f74 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.552483] env[63021]: DEBUG oslo_vmware.api [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Waiting for the task: (returnval){ [ 941.552483] env[63021]: value = "task-1294024" [ 941.552483] env[63021]: _type = "Task" [ 941.552483] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.561131] env[63021]: DEBUG oslo_vmware.api [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Task: {'id': task-1294024, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.595472] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.687947] env[63021]: DEBUG oslo_vmware.api [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294023, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.700464] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dd0e8b8d-f34d-42a3-9abe-8c123f54d6d9 tempest-ServerRescueTestJSON-1523849063 tempest-ServerRescueTestJSON-1523849063-project-member] Lock "df0b6974-7f78-4b73-8583-d09754fbacea" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.676s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.739834] env[63021]: DEBUG nova.compute.manager [req-d0e9c5c5-5a28-4503-93ca-88c780e34b1c req-a5eb04ff-b034-40b4-a3cf-edadd5cba439 service nova] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Received event network-vif-deleted-89471183-ed50-4a4d-85ce-790ebdf37c27 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 941.794021] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.808297] env[63021]: DEBUG nova.network.neutron [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 941.882509] env[63021]: DEBUG oslo_concurrency.lockutils [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.942861] env[63021]: DEBUG nova.network.neutron [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Updating instance_info_cache with network_info: [{"id": "45d08a72-44cf-4326-b881-7b60e5d8d962", "address": "fa:16:3e:71:de:2d", "network": {"id": "f2bef363-f8ef-4a12-8458-b4c0d71821b8", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1994701852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f659d3d817cc4d7c8d718e5374acb4cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ab9e5e6-9bf8-4a8d-91c8-d22148e3d2ee", "external-id": "nsx-vlan-transportzone-401", "segmentation_id": 401, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45d08a72-44", "ovs_interfaceid": "45d08a72-44cf-4326-b881-7b60e5d8d962", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.961929] env[63021]: INFO nova.compute.manager [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Rescuing [ 941.962239] env[63021]: DEBUG oslo_concurrency.lockutils [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "refresh_cache-ae103118-bddf-46fe-90b9-98b60952ebba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.962396] env[63021]: DEBUG oslo_concurrency.lockutils [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquired lock "refresh_cache-ae103118-bddf-46fe-90b9-98b60952ebba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.962562] env[63021]: DEBUG nova.network.neutron [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 942.062095] env[63021]: DEBUG oslo_vmware.api [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Task: {'id': task-1294024, 'name': ReconfigVM_Task, 'duration_secs': 0.340194} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.062563] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Reconfigured VM instance instance-0000004d to detach disk 2000 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 942.067013] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d64b80ec-917e-4a57-b9e3-7dcdab9a0430 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.083144] env[63021]: DEBUG oslo_vmware.api [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Waiting for the task: (returnval){ [ 942.083144] env[63021]: value = "task-1294025" [ 942.083144] env[63021]: _type = "Task" [ 942.083144] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.091154] env[63021]: DEBUG oslo_vmware.api [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Task: {'id': task-1294025, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.187354] env[63021]: DEBUG oslo_vmware.api [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294023, 'name': ReconfigVM_Task, 'duration_secs': 0.944613} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.187650] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Reconfigured VM instance instance-00000055 to attach disk [datastore2] 207c8197-6191-4837-8ead-8c24aa5b35bc/207c8197-6191-4837-8ead-8c24aa5b35bc.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 942.188332] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6b3c93f0-d12e-45dd-9ed2-f3fdbe3a419a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.194078] env[63021]: DEBUG oslo_vmware.api [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 942.194078] env[63021]: value = "task-1294026" [ 942.194078] env[63021]: _type = "Task" [ 942.194078] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.202318] env[63021]: DEBUG oslo_vmware.api [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294026, 'name': Rename_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.290457] env[63021]: DEBUG nova.compute.manager [req-79f5427e-7132-4aaf-81d1-0221919ff4af req-a5f91009-18f5-416a-82d3-46d9732da034 service nova] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Received event network-vif-plugged-45d08a72-44cf-4326-b881-7b60e5d8d962 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 942.290686] env[63021]: DEBUG oslo_concurrency.lockutils [req-79f5427e-7132-4aaf-81d1-0221919ff4af req-a5f91009-18f5-416a-82d3-46d9732da034 service nova] Acquiring lock "a52e4cb5-d0ea-4698-9955-753626762a02-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.290812] env[63021]: DEBUG oslo_concurrency.lockutils [req-79f5427e-7132-4aaf-81d1-0221919ff4af req-a5f91009-18f5-416a-82d3-46d9732da034 service nova] Lock "a52e4cb5-d0ea-4698-9955-753626762a02-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.291149] env[63021]: DEBUG oslo_concurrency.lockutils [req-79f5427e-7132-4aaf-81d1-0221919ff4af req-a5f91009-18f5-416a-82d3-46d9732da034 service nova] Lock "a52e4cb5-d0ea-4698-9955-753626762a02-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.291149] env[63021]: DEBUG nova.compute.manager [req-79f5427e-7132-4aaf-81d1-0221919ff4af req-a5f91009-18f5-416a-82d3-46d9732da034 service nova] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] No waiting events found dispatching network-vif-plugged-45d08a72-44cf-4326-b881-7b60e5d8d962 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 942.291485] env[63021]: WARNING nova.compute.manager [req-79f5427e-7132-4aaf-81d1-0221919ff4af req-a5f91009-18f5-416a-82d3-46d9732da034 service nova] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Received unexpected event network-vif-plugged-45d08a72-44cf-4326-b881-7b60e5d8d962 for instance with vm_state building and task_state spawning. [ 942.291485] env[63021]: DEBUG nova.compute.manager [req-79f5427e-7132-4aaf-81d1-0221919ff4af req-a5f91009-18f5-416a-82d3-46d9732da034 service nova] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Received event network-changed-45d08a72-44cf-4326-b881-7b60e5d8d962 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 942.291729] env[63021]: DEBUG nova.compute.manager [req-79f5427e-7132-4aaf-81d1-0221919ff4af req-a5f91009-18f5-416a-82d3-46d9732da034 service nova] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Refreshing instance network info cache due to event network-changed-45d08a72-44cf-4326-b881-7b60e5d8d962. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 942.291942] env[63021]: DEBUG oslo_concurrency.lockutils [req-79f5427e-7132-4aaf-81d1-0221919ff4af req-a5f91009-18f5-416a-82d3-46d9732da034 service nova] Acquiring lock "refresh_cache-a52e4cb5-d0ea-4698-9955-753626762a02" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.374870] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-813ddf1e-8e45-4b8c-bd5f-0c9b4e56cbf1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.382387] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb2d3c7e-5972-41dd-8e81-10fe1dad3ca5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.411990] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43006fb3-d868-4054-b747-b80aebbfac5b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.419220] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48c60506-906e-4895-be90-31537e5ec27e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.434405] env[63021]: DEBUG nova.compute.provider_tree [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 942.445248] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Releasing lock "refresh_cache-a52e4cb5-d0ea-4698-9955-753626762a02" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.445601] env[63021]: DEBUG nova.compute.manager [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Instance network_info: |[{"id": "45d08a72-44cf-4326-b881-7b60e5d8d962", "address": "fa:16:3e:71:de:2d", "network": {"id": "f2bef363-f8ef-4a12-8458-b4c0d71821b8", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1994701852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f659d3d817cc4d7c8d718e5374acb4cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ab9e5e6-9bf8-4a8d-91c8-d22148e3d2ee", "external-id": "nsx-vlan-transportzone-401", "segmentation_id": 401, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45d08a72-44", "ovs_interfaceid": "45d08a72-44cf-4326-b881-7b60e5d8d962", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 942.445938] env[63021]: DEBUG oslo_concurrency.lockutils [req-79f5427e-7132-4aaf-81d1-0221919ff4af req-a5f91009-18f5-416a-82d3-46d9732da034 service nova] Acquired lock "refresh_cache-a52e4cb5-d0ea-4698-9955-753626762a02" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.446152] env[63021]: DEBUG nova.network.neutron [req-79f5427e-7132-4aaf-81d1-0221919ff4af req-a5f91009-18f5-416a-82d3-46d9732da034 service nova] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Refreshing network info cache for port 45d08a72-44cf-4326-b881-7b60e5d8d962 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 942.447300] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:71:de:2d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8ab9e5e6-9bf8-4a8d-91c8-d22148e3d2ee', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '45d08a72-44cf-4326-b881-7b60e5d8d962', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 942.455703] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Creating folder: Project (f659d3d817cc4d7c8d718e5374acb4cb). Parent ref: group-v277447. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 942.456438] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6e4cfafa-43ba-4dec-9665-b1d56e9c6b7b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.468291] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Created folder: Project (f659d3d817cc4d7c8d718e5374acb4cb) in parent group-v277447. [ 942.468478] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Creating folder: Instances. Parent ref: group-v277558. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 942.468993] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fdc227b9-b669-496c-a675-7aac9a553b5e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.477148] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Created folder: Instances in parent group-v277558. [ 942.477392] env[63021]: DEBUG oslo.service.loopingcall [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 942.477579] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 942.477814] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-196bc1d5-a207-43a1-b221-228d7e7db75d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.496091] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 942.496091] env[63021]: value = "task-1294029" [ 942.496091] env[63021]: _type = "Task" [ 942.496091] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.505985] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294029, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.594740] env[63021]: DEBUG oslo_vmware.api [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Task: {'id': task-1294025, 'name': ReconfigVM_Task, 'duration_secs': 0.106543} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.595080] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277493', 'volume_id': 'e424adc5-a8b0-4409-b3e0-f3cdbc8fb712', 'name': 'volume-e424adc5-a8b0-4409-b3e0-f3cdbc8fb712', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '80b857c4-d9dd-4483-970f-a0f5f4bcf173', 'attached_at': '', 'detached_at': '', 'volume_id': 'e424adc5-a8b0-4409-b3e0-f3cdbc8fb712', 'serial': 'e424adc5-a8b0-4409-b3e0-f3cdbc8fb712'} {{(pid=63021) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 942.595442] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 942.596327] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-349b60b6-f856-466c-8e30-0f58007f30e5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.603453] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 942.603767] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b2e093dd-36da-4768-b3bb-af37b329dd10 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.700145] env[63021]: DEBUG nova.network.neutron [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Updating instance_info_cache with network_info: [{"id": "f99a201d-df02-499f-85d9-2566ac1dd82b", "address": "fa:16:3e:a0:fd:ff", "network": {"id": "3aaeed35-f452-4144-a259-8f440f8a8bf4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-229941363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b672b1d7d7ad454ebfa3e2fd2421c754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf99a201d-df", "ovs_interfaceid": "f99a201d-df02-499f-85d9-2566ac1dd82b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.706404] env[63021]: DEBUG oslo_vmware.api [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294026, 'name': Rename_Task, 'duration_secs': 0.410445} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.706683] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 942.706939] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c8a119bb-b6b6-487a-9fa3-0f38c0984532 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.714151] env[63021]: DEBUG oslo_vmware.api [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 942.714151] env[63021]: value = "task-1294031" [ 942.714151] env[63021]: _type = "Task" [ 942.714151] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.722878] env[63021]: DEBUG oslo_vmware.api [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294031, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.808224] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 942.808224] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Deleting contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 942.808224] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Deleting the datastore file [datastore2] 80b857c4-d9dd-4483-970f-a0f5f4bcf173 {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 942.808224] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c3628f0c-fcfd-4d74-b609-e876bf02a29e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.812846] env[63021]: DEBUG oslo_vmware.api [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Waiting for the task: (returnval){ [ 942.812846] env[63021]: value = "task-1294032" [ 942.812846] env[63021]: _type = "Task" [ 942.812846] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.821316] env[63021]: DEBUG oslo_vmware.api [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Task: {'id': task-1294032, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.938163] env[63021]: DEBUG nova.scheduler.client.report [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 943.008546] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294029, 'name': CreateVM_Task, 'duration_secs': 0.427902} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.008724] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 943.009424] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.009596] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.009957] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 943.010236] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd766cea-4fd2-416a-ad94-f633e395d967 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.014754] env[63021]: DEBUG oslo_vmware.api [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Waiting for the task: (returnval){ [ 943.014754] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52ea5298-b2fe-8eef-6e8d-d8f745d09c19" [ 943.014754] env[63021]: _type = "Task" [ 943.014754] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.022683] env[63021]: DEBUG oslo_vmware.api [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52ea5298-b2fe-8eef-6e8d-d8f745d09c19, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.203289] env[63021]: DEBUG oslo_concurrency.lockutils [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Releasing lock "refresh_cache-ae103118-bddf-46fe-90b9-98b60952ebba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.206484] env[63021]: DEBUG nova.network.neutron [req-79f5427e-7132-4aaf-81d1-0221919ff4af req-a5f91009-18f5-416a-82d3-46d9732da034 service nova] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Updated VIF entry in instance network info cache for port 45d08a72-44cf-4326-b881-7b60e5d8d962. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 943.206891] env[63021]: DEBUG nova.network.neutron [req-79f5427e-7132-4aaf-81d1-0221919ff4af req-a5f91009-18f5-416a-82d3-46d9732da034 service nova] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Updating instance_info_cache with network_info: [{"id": "45d08a72-44cf-4326-b881-7b60e5d8d962", "address": "fa:16:3e:71:de:2d", "network": {"id": "f2bef363-f8ef-4a12-8458-b4c0d71821b8", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1994701852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f659d3d817cc4d7c8d718e5374acb4cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ab9e5e6-9bf8-4a8d-91c8-d22148e3d2ee", "external-id": "nsx-vlan-transportzone-401", "segmentation_id": 401, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45d08a72-44", "ovs_interfaceid": "45d08a72-44cf-4326-b881-7b60e5d8d962", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.228360] env[63021]: DEBUG oslo_vmware.api [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294031, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.322745] env[63021]: DEBUG oslo_vmware.api [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Task: {'id': task-1294032, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.163555} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.323441] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 943.323861] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Deleted contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 943.323861] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 943.380281] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Volume detach. Driver type: vmdk {{(pid=63021) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 943.380652] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cd9db4ef-da79-439e-8b4a-9e5a29b000c9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.392259] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57c50a22-c28d-43c1-89db-07f829067ab9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.420631] env[63021]: ERROR nova.compute.manager [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Failed to detach volume e424adc5-a8b0-4409-b3e0-f3cdbc8fb712 from /dev/sda: nova.exception.InstanceNotFound: Instance 80b857c4-d9dd-4483-970f-a0f5f4bcf173 could not be found. [ 943.420631] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Traceback (most recent call last): [ 943.420631] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 943.420631] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] self.driver.rebuild(**kwargs) [ 943.420631] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 943.420631] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] raise NotImplementedError() [ 943.420631] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] NotImplementedError [ 943.420631] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] [ 943.420631] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] During handling of the above exception, another exception occurred: [ 943.420631] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] [ 943.420631] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Traceback (most recent call last): [ 943.420631] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 943.420631] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] self.driver.detach_volume(context, old_connection_info, [ 943.420631] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 553, in detach_volume [ 943.420631] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] return self._volumeops.detach_volume(connection_info, instance) [ 943.420631] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 943.420631] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] self._detach_volume_vmdk(connection_info, instance) [ 943.420631] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 943.420631] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 943.420631] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 943.420631] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] stable_ref.fetch_moref(session) [ 943.420631] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 943.420631] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] raise exception.InstanceNotFound(instance_id=self._uuid) [ 943.420631] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] nova.exception.InstanceNotFound: Instance 80b857c4-d9dd-4483-970f-a0f5f4bcf173 could not be found. [ 943.420631] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] [ 943.443628] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.276s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.444191] env[63021]: DEBUG nova.compute.manager [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 943.449252] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.276s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.449484] env[63021]: DEBUG nova.objects.instance [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lazy-loading 'resources' on Instance uuid 881c2fb8-5f8c-48c2-a173-8fab949f24f7 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 943.529103] env[63021]: DEBUG oslo_vmware.api [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52ea5298-b2fe-8eef-6e8d-d8f745d09c19, 'name': SearchDatastore_Task, 'duration_secs': 0.014158} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.529600] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.530495] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 943.530495] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.530798] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.531070] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 943.531398] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-edf77791-abe7-475c-9c7d-2de0646e9857 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.543018] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 943.543280] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 943.543996] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c156d62f-d9a4-433c-b589-164929ede120 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.551012] env[63021]: DEBUG oslo_vmware.api [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Waiting for the task: (returnval){ [ 943.551012] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5211d9a0-196e-c4f0-70b6-d435b8a6d9bd" [ 943.551012] env[63021]: _type = "Task" [ 943.551012] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.558706] env[63021]: DEBUG oslo_vmware.api [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5211d9a0-196e-c4f0-70b6-d435b8a6d9bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.578920] env[63021]: DEBUG nova.compute.utils [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Build of instance 80b857c4-d9dd-4483-970f-a0f5f4bcf173 aborted: Failed to rebuild volume backed instance. {{(pid=63021) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 943.581077] env[63021]: ERROR nova.compute.manager [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance 80b857c4-d9dd-4483-970f-a0f5f4bcf173 aborted: Failed to rebuild volume backed instance. [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Traceback (most recent call last): [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] self.driver.rebuild(**kwargs) [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] raise NotImplementedError() [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] NotImplementedError [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] During handling of the above exception, another exception occurred: [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Traceback (most recent call last): [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] File "/opt/stack/nova/nova/compute/manager.py", line 3600, in _rebuild_volume_backed_instance [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] self._detach_root_volume(context, instance, root_bdm) [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] File "/opt/stack/nova/nova/compute/manager.py", line 3579, in _detach_root_volume [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] with excutils.save_and_reraise_exception(): [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] self.force_reraise() [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] raise self.value [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] self.driver.detach_volume(context, old_connection_info, [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 553, in detach_volume [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] return self._volumeops.detach_volume(connection_info, instance) [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] self._detach_volume_vmdk(connection_info, instance) [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] stable_ref.fetch_moref(session) [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] raise exception.InstanceNotFound(instance_id=self._uuid) [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] nova.exception.InstanceNotFound: Instance 80b857c4-d9dd-4483-970f-a0f5f4bcf173 could not be found. [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] During handling of the above exception, another exception occurred: [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Traceback (most recent call last): [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] File "/opt/stack/nova/nova/compute/manager.py", line 10865, in _error_out_instance_on_exception [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] yield [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] File "/opt/stack/nova/nova/compute/manager.py", line 3868, in rebuild_instance [ 943.581077] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] self._do_rebuild_instance_with_claim( [ 943.582460] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] File "/opt/stack/nova/nova/compute/manager.py", line 3954, in _do_rebuild_instance_with_claim [ 943.582460] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] self._do_rebuild_instance( [ 943.582460] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] File "/opt/stack/nova/nova/compute/manager.py", line 4146, in _do_rebuild_instance [ 943.582460] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] self._rebuild_default_impl(**kwargs) [ 943.582460] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] File "/opt/stack/nova/nova/compute/manager.py", line 3723, in _rebuild_default_impl [ 943.582460] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] self._rebuild_volume_backed_instance( [ 943.582460] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] File "/opt/stack/nova/nova/compute/manager.py", line 3615, in _rebuild_volume_backed_instance [ 943.582460] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] raise exception.BuildAbortException( [ 943.582460] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] nova.exception.BuildAbortException: Build of instance 80b857c4-d9dd-4483-970f-a0f5f4bcf173 aborted: Failed to rebuild volume backed instance. [ 943.582460] env[63021]: ERROR nova.compute.manager [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] [ 943.708993] env[63021]: DEBUG oslo_concurrency.lockutils [req-79f5427e-7132-4aaf-81d1-0221919ff4af req-a5f91009-18f5-416a-82d3-46d9732da034 service nova] Releasing lock "refresh_cache-a52e4cb5-d0ea-4698-9955-753626762a02" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.725875] env[63021]: DEBUG oslo_vmware.api [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294031, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.735217] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 943.735527] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3f33ff92-7b3a-442c-ae41-284ca7b8fa7e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.741436] env[63021]: DEBUG oslo_vmware.api [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 943.741436] env[63021]: value = "task-1294033" [ 943.741436] env[63021]: _type = "Task" [ 943.741436] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.749307] env[63021]: DEBUG oslo_vmware.api [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294033, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.952901] env[63021]: DEBUG nova.compute.utils [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 943.957393] env[63021]: DEBUG nova.compute.manager [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 943.957488] env[63021]: DEBUG nova.network.neutron [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 944.004596] env[63021]: DEBUG nova.policy [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1e133aeda3249ec98dbdc53f869d719', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7331caed42034dfc81c79cfabbe0829e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 944.064573] env[63021]: DEBUG oslo_vmware.api [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5211d9a0-196e-c4f0-70b6-d435b8a6d9bd, 'name': SearchDatastore_Task, 'duration_secs': 0.008331} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.065924] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-23ed99ea-2257-4fab-a636-7e631e790be2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.073117] env[63021]: DEBUG oslo_vmware.api [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Waiting for the task: (returnval){ [ 944.073117] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52d62d3c-fdcb-6dd0-ebe2-31e012ba7527" [ 944.073117] env[63021]: _type = "Task" [ 944.073117] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.081124] env[63021]: DEBUG oslo_vmware.api [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52d62d3c-fdcb-6dd0-ebe2-31e012ba7527, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.132086] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5529e1cf-8df1-48a3-8f3f-4f554378bb7d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.142951] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46405651-d592-425a-afd6-e807d1cb1921 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.174411] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3b471e2-210a-4f98-8b30-d9f9dbf55e52 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.182365] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbf838f4-74b9-454a-aac6-4a229090b3e5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.196676] env[63021]: DEBUG nova.compute.provider_tree [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 944.230021] env[63021]: DEBUG oslo_vmware.api [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294031, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.251595] env[63021]: DEBUG oslo_vmware.api [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294033, 'name': PowerOffVM_Task, 'duration_secs': 0.164875} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.251899] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 944.252784] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5514c31b-7948-4c30-a19d-693199729243 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.276363] env[63021]: DEBUG nova.network.neutron [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Successfully created port: bb825b85-647f-45b0-b351-ec7921e3ae5e {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 944.279059] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19bce623-b8c5-4871-80ce-c52a0cd83475 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.315718] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 944.316059] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8950595f-115e-4d2c-b0e9-5501b98c0eb1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.322774] env[63021]: DEBUG oslo_vmware.api [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 944.322774] env[63021]: value = "task-1294034" [ 944.322774] env[63021]: _type = "Task" [ 944.322774] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.329891] env[63021]: DEBUG oslo_vmware.api [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294034, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.458671] env[63021]: DEBUG nova.compute.manager [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 944.584537] env[63021]: DEBUG oslo_vmware.api [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52d62d3c-fdcb-6dd0-ebe2-31e012ba7527, 'name': SearchDatastore_Task, 'duration_secs': 0.010113} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.585085] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.585391] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] a52e4cb5-d0ea-4698-9955-753626762a02/a52e4cb5-d0ea-4698-9955-753626762a02.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 944.585701] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-27cc4c24-fe31-47fb-833f-569b64984c03 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.598187] env[63021]: DEBUG oslo_vmware.api [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Waiting for the task: (returnval){ [ 944.598187] env[63021]: value = "task-1294035" [ 944.598187] env[63021]: _type = "Task" [ 944.598187] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.604074] env[63021]: DEBUG oslo_vmware.api [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Task: {'id': task-1294035, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.702021] env[63021]: DEBUG nova.scheduler.client.report [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 944.727391] env[63021]: DEBUG oslo_vmware.api [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294031, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.838381] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] VM already powered off {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 944.838381] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 944.838381] env[63021]: DEBUG oslo_concurrency.lockutils [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.838381] env[63021]: DEBUG oslo_concurrency.lockutils [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.838381] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 944.838381] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-02a9e5d5-1da8-4e85-815c-07282ad4458d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.853493] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 944.853493] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 944.853493] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e08beac-0787-429f-b99d-fce8d146a85a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.863293] env[63021]: DEBUG oslo_vmware.api [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 944.863293] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52df2a44-11f8-4ca3-957a-a356bbcc35ab" [ 944.863293] env[63021]: _type = "Task" [ 944.863293] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.873658] env[63021]: DEBUG oslo_vmware.api [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52df2a44-11f8-4ca3-957a-a356bbcc35ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.110084] env[63021]: DEBUG oslo_vmware.api [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Task: {'id': task-1294035, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.478435} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.110084] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] a52e4cb5-d0ea-4698-9955-753626762a02/a52e4cb5-d0ea-4698-9955-753626762a02.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 945.110084] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 945.110084] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5a442fd1-68c2-4942-a6ec-a7d2c2b8bdeb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.116384] env[63021]: DEBUG oslo_vmware.api [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Waiting for the task: (returnval){ [ 945.116384] env[63021]: value = "task-1294036" [ 945.116384] env[63021]: _type = "Task" [ 945.116384] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.123104] env[63021]: DEBUG oslo_vmware.api [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Task: {'id': task-1294036, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.206498] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.757s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.210341] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.613s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.210341] env[63021]: DEBUG nova.objects.instance [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lazy-loading 'resources' on Instance uuid 140184cd-6e4c-425c-8b17-361a1f565e93 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 945.234024] env[63021]: DEBUG oslo_vmware.api [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294031, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.239562] env[63021]: INFO nova.scheduler.client.report [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Deleted allocations for instance 881c2fb8-5f8c-48c2-a173-8fab949f24f7 [ 945.376811] env[63021]: DEBUG oslo_vmware.api [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52df2a44-11f8-4ca3-957a-a356bbcc35ab, 'name': SearchDatastore_Task, 'duration_secs': 0.023356} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.378078] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a9e147b-15dd-4bb3-a303-3e22da1c70b4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.383620] env[63021]: DEBUG oslo_vmware.api [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 945.383620] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52761a60-0fd0-a209-43eb-971c8b9da051" [ 945.383620] env[63021]: _type = "Task" [ 945.383620] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.394204] env[63021]: DEBUG oslo_vmware.api [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52761a60-0fd0-a209-43eb-971c8b9da051, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.469549] env[63021]: DEBUG nova.compute.manager [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 945.499140] env[63021]: DEBUG nova.virt.hardware [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 945.499481] env[63021]: DEBUG nova.virt.hardware [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 945.499698] env[63021]: DEBUG nova.virt.hardware [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 945.499962] env[63021]: DEBUG nova.virt.hardware [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 945.500186] env[63021]: DEBUG nova.virt.hardware [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 945.500403] env[63021]: DEBUG nova.virt.hardware [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 945.500739] env[63021]: DEBUG nova.virt.hardware [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 945.500984] env[63021]: DEBUG nova.virt.hardware [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 945.501237] env[63021]: DEBUG nova.virt.hardware [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 945.501465] env[63021]: DEBUG nova.virt.hardware [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 945.501709] env[63021]: DEBUG nova.virt.hardware [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 945.502880] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50f73a2e-b95e-4357-bde0-d1516115b153 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.513145] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05948990-c301-45fc-bbe2-f7a21bbcc241 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.603906] env[63021]: DEBUG oslo_concurrency.lockutils [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.629634] env[63021]: DEBUG oslo_vmware.api [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Task: {'id': task-1294036, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061976} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.629634] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 945.629634] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c77b06e-e0fb-4866-b951-22a95f55abc8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.654521] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] a52e4cb5-d0ea-4698-9955-753626762a02/a52e4cb5-d0ea-4698-9955-753626762a02.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 945.657160] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dc13ebde-786d-4e0d-9647-c1eace093f61 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.680817] env[63021]: DEBUG oslo_vmware.api [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Waiting for the task: (returnval){ [ 945.680817] env[63021]: value = "task-1294037" [ 945.680817] env[63021]: _type = "Task" [ 945.680817] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.692644] env[63021]: DEBUG oslo_vmware.api [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Task: {'id': task-1294037, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.694816] env[63021]: DEBUG nova.compute.manager [req-9c748024-5e51-4e32-9053-50caa7692948 req-b7fdf81a-f6db-4db4-bef1-471f40c58236 service nova] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Received event network-vif-plugged-bb825b85-647f-45b0-b351-ec7921e3ae5e {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 945.695116] env[63021]: DEBUG oslo_concurrency.lockutils [req-9c748024-5e51-4e32-9053-50caa7692948 req-b7fdf81a-f6db-4db4-bef1-471f40c58236 service nova] Acquiring lock "919887c8-171d-48fe-8c6b-2abf5f55d6ab-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.695501] env[63021]: DEBUG oslo_concurrency.lockutils [req-9c748024-5e51-4e32-9053-50caa7692948 req-b7fdf81a-f6db-4db4-bef1-471f40c58236 service nova] Lock "919887c8-171d-48fe-8c6b-2abf5f55d6ab-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.695711] env[63021]: DEBUG oslo_concurrency.lockutils [req-9c748024-5e51-4e32-9053-50caa7692948 req-b7fdf81a-f6db-4db4-bef1-471f40c58236 service nova] Lock "919887c8-171d-48fe-8c6b-2abf5f55d6ab-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.695957] env[63021]: DEBUG nova.compute.manager [req-9c748024-5e51-4e32-9053-50caa7692948 req-b7fdf81a-f6db-4db4-bef1-471f40c58236 service nova] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] No waiting events found dispatching network-vif-plugged-bb825b85-647f-45b0-b351-ec7921e3ae5e {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 945.696234] env[63021]: WARNING nova.compute.manager [req-9c748024-5e51-4e32-9053-50caa7692948 req-b7fdf81a-f6db-4db4-bef1-471f40c58236 service nova] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Received unexpected event network-vif-plugged-bb825b85-647f-45b0-b351-ec7921e3ae5e for instance with vm_state building and task_state spawning. [ 945.728503] env[63021]: DEBUG oslo_vmware.api [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294031, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.748613] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3a54c3d9-5cf6-4dbf-bcfe-e246180c328d tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "881c2fb8-5f8c-48c2-a173-8fab949f24f7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.610s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.771316] env[63021]: DEBUG nova.network.neutron [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Successfully updated port: bb825b85-647f-45b0-b351-ec7921e3ae5e {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 945.896082] env[63021]: DEBUG oslo_vmware.api [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52761a60-0fd0-a209-43eb-971c8b9da051, 'name': SearchDatastore_Task, 'duration_secs': 0.026538} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.896363] env[63021]: DEBUG oslo_concurrency.lockutils [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.896621] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] ae103118-bddf-46fe-90b9-98b60952ebba/e624225d-af0f-415c-90cc-d4b228c479f6-rescue.vmdk. {{(pid=63021) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 945.896978] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1d93b370-0689-4da8-99e9-dd0b6a97cbb7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.903453] env[63021]: DEBUG oslo_vmware.api [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 945.903453] env[63021]: value = "task-1294038" [ 945.903453] env[63021]: _type = "Task" [ 945.903453] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.914177] env[63021]: DEBUG oslo_vmware.api [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294038, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.922333] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b977d118-7c35-4742-8dfd-ab4a3242f22e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.929016] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54b61a6f-163d-4082-a5fe-59cb80b12eb0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.963652] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d10156c-f357-4c10-b635-7271577717bf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.971700] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fb18642-11b6-4146-9160-1b003b2d1492 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.985935] env[63021]: DEBUG nova.compute.provider_tree [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 946.192832] env[63021]: DEBUG oslo_vmware.api [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Task: {'id': task-1294037, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.230324] env[63021]: DEBUG oslo_vmware.api [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294031, 'name': PowerOnVM_Task} progress is 74%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.273077] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "refresh_cache-919887c8-171d-48fe-8c6b-2abf5f55d6ab" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.273077] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquired lock "refresh_cache-919887c8-171d-48fe-8c6b-2abf5f55d6ab" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.273662] env[63021]: DEBUG nova.network.neutron [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 946.416766] env[63021]: DEBUG oslo_vmware.api [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294038, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.489350] env[63021]: DEBUG nova.scheduler.client.report [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 946.639732] env[63021]: DEBUG oslo_concurrency.lockutils [None req-db9db38f-ae08-4a7b-8cb1-5d2701733145 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Acquiring lock "80b857c4-d9dd-4483-970f-a0f5f4bcf173" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.640127] env[63021]: DEBUG oslo_concurrency.lockutils [None req-db9db38f-ae08-4a7b-8cb1-5d2701733145 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Lock "80b857c4-d9dd-4483-970f-a0f5f4bcf173" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.640423] env[63021]: DEBUG oslo_concurrency.lockutils [None req-db9db38f-ae08-4a7b-8cb1-5d2701733145 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Acquiring lock "80b857c4-d9dd-4483-970f-a0f5f4bcf173-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.640560] env[63021]: DEBUG oslo_concurrency.lockutils [None req-db9db38f-ae08-4a7b-8cb1-5d2701733145 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Lock "80b857c4-d9dd-4483-970f-a0f5f4bcf173-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.640728] env[63021]: DEBUG oslo_concurrency.lockutils [None req-db9db38f-ae08-4a7b-8cb1-5d2701733145 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Lock "80b857c4-d9dd-4483-970f-a0f5f4bcf173-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.642856] env[63021]: INFO nova.compute.manager [None req-db9db38f-ae08-4a7b-8cb1-5d2701733145 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Terminating instance [ 946.646087] env[63021]: DEBUG nova.compute.manager [None req-db9db38f-ae08-4a7b-8cb1-5d2701733145 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 946.646253] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0f37e9b3-2faa-46f5-b6bc-03c33155d143 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.654979] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0624a90d-ea8a-43ea-8290-6fbbea59e5e5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.685814] env[63021]: WARNING nova.virt.vmwareapi.driver [None req-db9db38f-ae08-4a7b-8cb1-5d2701733145 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 80b857c4-d9dd-4483-970f-a0f5f4bcf173 could not be found. [ 946.686073] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-db9db38f-ae08-4a7b-8cb1-5d2701733145 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 946.689422] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e91f41a0-3d34-4fe4-ba64-8ded5e92bbd1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.696705] env[63021]: DEBUG oslo_vmware.api [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Task: {'id': task-1294037, 'name': ReconfigVM_Task, 'duration_secs': 0.645503} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.698010] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Reconfigured VM instance instance-00000056 to attach disk [datastore2] a52e4cb5-d0ea-4698-9955-753626762a02/a52e4cb5-d0ea-4698-9955-753626762a02.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 946.699644] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8e41da47-0000-4e62-9786-089790744a68 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.703056] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-526b01fd-2473-4095-ab4f-61e78f83d03f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.720838] env[63021]: DEBUG oslo_vmware.api [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Waiting for the task: (returnval){ [ 946.720838] env[63021]: value = "task-1294039" [ 946.720838] env[63021]: _type = "Task" [ 946.720838] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.740515] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-db9db38f-ae08-4a7b-8cb1-5d2701733145 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 80b857c4-d9dd-4483-970f-a0f5f4bcf173 could not be found. [ 946.740654] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-db9db38f-ae08-4a7b-8cb1-5d2701733145 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 946.740915] env[63021]: INFO nova.compute.manager [None req-db9db38f-ae08-4a7b-8cb1-5d2701733145 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Took 0.09 seconds to destroy the instance on the hypervisor. [ 946.741251] env[63021]: DEBUG oslo.service.loopingcall [None req-db9db38f-ae08-4a7b-8cb1-5d2701733145 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 946.742524] env[63021]: DEBUG nova.compute.manager [-] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 946.742649] env[63021]: DEBUG nova.network.neutron [-] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 946.750679] env[63021]: DEBUG oslo_vmware.api [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294031, 'name': PowerOnVM_Task, 'duration_secs': 3.809572} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.753138] env[63021]: DEBUG oslo_vmware.api [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Task: {'id': task-1294039, 'name': Rename_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.753138] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 946.753138] env[63021]: INFO nova.compute.manager [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Took 10.94 seconds to spawn the instance on the hypervisor. [ 946.753138] env[63021]: DEBUG nova.compute.manager [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 946.753138] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c4c5e70-e378-4826-a7ff-035cb9edef35 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.847335] env[63021]: DEBUG nova.network.neutron [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 946.919052] env[63021]: DEBUG oslo_vmware.api [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294038, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.564487} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.919052] env[63021]: INFO nova.virt.vmwareapi.ds_util [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] ae103118-bddf-46fe-90b9-98b60952ebba/e624225d-af0f-415c-90cc-d4b228c479f6-rescue.vmdk. [ 946.919141] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f717f77a-b24e-4eee-a133-0dd92b6e3540 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.947081] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] ae103118-bddf-46fe-90b9-98b60952ebba/e624225d-af0f-415c-90cc-d4b228c479f6-rescue.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 946.947448] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-056b18ec-be40-4807-b3d4-f167a025d102 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.966334] env[63021]: DEBUG oslo_vmware.api [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 946.966334] env[63021]: value = "task-1294040" [ 946.966334] env[63021]: _type = "Task" [ 946.966334] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.974854] env[63021]: DEBUG oslo_vmware.api [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294040, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.999012] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.790s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.001854] env[63021]: DEBUG nova.network.neutron [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Updating instance_info_cache with network_info: [{"id": "bb825b85-647f-45b0-b351-ec7921e3ae5e", "address": "fa:16:3e:fd:c3:4f", "network": {"id": "265e2e2e-6ec6-4d4c-8b4c-0a2e8baa6bdb", "bridge": "br-int", "label": "tempest-ImagesTestJSON-242691215-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7331caed42034dfc81c79cfabbe0829e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bfae3ef8-cae7-455d-8632-ba93e1671625", "external-id": "cl2-zone-841", "segmentation_id": 841, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb825b85-64", "ovs_interfaceid": "bb825b85-647f-45b0-b351-ec7921e3ae5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.006017] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.210s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.006017] env[63021]: INFO nova.compute.claims [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 947.045224] env[63021]: INFO nova.scheduler.client.report [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Deleted allocations for instance 140184cd-6e4c-425c-8b17-361a1f565e93 [ 947.231839] env[63021]: DEBUG oslo_vmware.api [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Task: {'id': task-1294039, 'name': Rename_Task, 'duration_secs': 0.385721} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.232155] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 947.232417] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8c9f54ba-b7d6-4c35-b7c4-c756150652d9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.238205] env[63021]: DEBUG oslo_vmware.api [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Waiting for the task: (returnval){ [ 947.238205] env[63021]: value = "task-1294041" [ 947.238205] env[63021]: _type = "Task" [ 947.238205] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.245978] env[63021]: DEBUG oslo_vmware.api [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Task: {'id': task-1294041, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.272726] env[63021]: INFO nova.compute.manager [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Took 17.41 seconds to build instance. [ 947.484128] env[63021]: DEBUG oslo_vmware.api [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294040, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.511448] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Releasing lock "refresh_cache-919887c8-171d-48fe-8c6b-2abf5f55d6ab" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.513962] env[63021]: DEBUG nova.compute.manager [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Instance network_info: |[{"id": "bb825b85-647f-45b0-b351-ec7921e3ae5e", "address": "fa:16:3e:fd:c3:4f", "network": {"id": "265e2e2e-6ec6-4d4c-8b4c-0a2e8baa6bdb", "bridge": "br-int", "label": "tempest-ImagesTestJSON-242691215-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7331caed42034dfc81c79cfabbe0829e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bfae3ef8-cae7-455d-8632-ba93e1671625", "external-id": "cl2-zone-841", "segmentation_id": 841, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb825b85-64", "ovs_interfaceid": "bb825b85-647f-45b0-b351-ec7921e3ae5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 947.516105] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fd:c3:4f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bfae3ef8-cae7-455d-8632-ba93e1671625', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bb825b85-647f-45b0-b351-ec7921e3ae5e', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 947.527801] env[63021]: DEBUG oslo.service.loopingcall [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 947.528023] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 947.529698] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-761d8365-e2c4-4757-86fc-d6a7b322652e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.557052] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 947.557052] env[63021]: value = "task-1294042" [ 947.557052] env[63021]: _type = "Task" [ 947.557052] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.557369] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0d4db7f-d049-47df-b4d4-feb194cbb3af tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "140184cd-6e4c-425c-8b17-361a1f565e93" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.832s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.571252] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294042, 'name': CreateVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.726888] env[63021]: DEBUG nova.compute.manager [req-d01e3e11-d605-45af-b3ff-825fcd03ad2b req-ae0cfa25-5b37-4a8f-b9af-815f1d3a4435 service nova] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Received event network-changed-bb825b85-647f-45b0-b351-ec7921e3ae5e {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 947.727996] env[63021]: DEBUG nova.compute.manager [req-d01e3e11-d605-45af-b3ff-825fcd03ad2b req-ae0cfa25-5b37-4a8f-b9af-815f1d3a4435 service nova] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Refreshing instance network info cache due to event network-changed-bb825b85-647f-45b0-b351-ec7921e3ae5e. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 947.728264] env[63021]: DEBUG oslo_concurrency.lockutils [req-d01e3e11-d605-45af-b3ff-825fcd03ad2b req-ae0cfa25-5b37-4a8f-b9af-815f1d3a4435 service nova] Acquiring lock "refresh_cache-919887c8-171d-48fe-8c6b-2abf5f55d6ab" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.728416] env[63021]: DEBUG oslo_concurrency.lockutils [req-d01e3e11-d605-45af-b3ff-825fcd03ad2b req-ae0cfa25-5b37-4a8f-b9af-815f1d3a4435 service nova] Acquired lock "refresh_cache-919887c8-171d-48fe-8c6b-2abf5f55d6ab" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.728658] env[63021]: DEBUG nova.network.neutron [req-d01e3e11-d605-45af-b3ff-825fcd03ad2b req-ae0cfa25-5b37-4a8f-b9af-815f1d3a4435 service nova] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Refreshing network info cache for port bb825b85-647f-45b0-b351-ec7921e3ae5e {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 947.750792] env[63021]: DEBUG oslo_vmware.api [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Task: {'id': task-1294041, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.775487] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3ee25e80-ca0f-469e-96f6-f1005daeafed tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "207c8197-6191-4837-8ead-8c24aa5b35bc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.925s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.816417] env[63021]: DEBUG nova.network.neutron [-] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.982441] env[63021]: DEBUG oslo_vmware.api [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294040, 'name': ReconfigVM_Task, 'duration_secs': 0.679107} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.982770] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Reconfigured VM instance instance-00000052 to attach disk [datastore1] ae103118-bddf-46fe-90b9-98b60952ebba/e624225d-af0f-415c-90cc-d4b228c479f6-rescue.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 947.984798] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39ca4df6-f4d6-494f-9392-933ec2e9e701 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.014480] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-86f59eb3-7272-4e9f-9749-2290af1aff39 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.036683] env[63021]: DEBUG oslo_vmware.api [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 948.036683] env[63021]: value = "task-1294043" [ 948.036683] env[63021]: _type = "Task" [ 948.036683] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.045466] env[63021]: DEBUG oslo_vmware.api [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294043, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.072898] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294042, 'name': CreateVM_Task, 'duration_secs': 0.374293} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.072898] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 948.072898] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.072898] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.073273] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 948.073314] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1492e1ce-22fe-4e1f-aebc-495f0c28cbf5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.079082] env[63021]: DEBUG oslo_vmware.api [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 948.079082] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52a1552f-7793-04ea-66a2-64e5f2ee8130" [ 948.079082] env[63021]: _type = "Task" [ 948.079082] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.091027] env[63021]: DEBUG oslo_vmware.api [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52a1552f-7793-04ea-66a2-64e5f2ee8130, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.149820] env[63021]: DEBUG oslo_concurrency.lockutils [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Acquiring lock "7df37eb7-ffc1-4b8e-9aba-caaf3c5338df" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.149820] env[63021]: DEBUG oslo_concurrency.lockutils [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Lock "7df37eb7-ffc1-4b8e-9aba-caaf3c5338df" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.250812] env[63021]: DEBUG oslo_vmware.api [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Task: {'id': task-1294041, 'name': PowerOnVM_Task, 'duration_secs': 0.693496} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.254373] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 948.254794] env[63021]: INFO nova.compute.manager [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Took 8.21 seconds to spawn the instance on the hypervisor. [ 948.255077] env[63021]: DEBUG nova.compute.manager [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 948.256669] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aab8bfb-5c71-4247-96a4-ff414df81524 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.286660] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1729cab7-5155-409a-b816-10446e772870 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.296092] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10c5901e-f3ea-4d51-a4fc-4dff7e154763 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.336217] env[63021]: INFO nova.compute.manager [-] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Took 1.59 seconds to deallocate network for instance. [ 948.340292] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0687e2f5-23bf-4be4-b37a-0b1b76433246 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.349851] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42ee198b-e315-4714-bb37-87ea962685f5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.366360] env[63021]: DEBUG nova.compute.provider_tree [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 948.549887] env[63021]: DEBUG oslo_vmware.api [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294043, 'name': ReconfigVM_Task, 'duration_secs': 0.275182} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.549887] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 948.549887] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ce01c4bd-19c3-4890-a63b-cf5e97d83f1b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.555469] env[63021]: DEBUG oslo_vmware.api [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 948.555469] env[63021]: value = "task-1294044" [ 948.555469] env[63021]: _type = "Task" [ 948.555469] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.566197] env[63021]: DEBUG oslo_vmware.api [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294044, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.592074] env[63021]: DEBUG oslo_vmware.api [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52a1552f-7793-04ea-66a2-64e5f2ee8130, 'name': SearchDatastore_Task, 'duration_secs': 0.010249} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.592327] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.592567] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 948.592814] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.592937] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.593127] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 948.593388] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6f95c3df-960e-45b3-a0fc-85760a1cda3a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.602296] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 948.602475] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 948.603239] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96add477-f15a-434d-9447-1c86679e015e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.609341] env[63021]: DEBUG oslo_vmware.api [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 948.609341] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]520fe007-e5f0-67ab-5b54-dbc133537304" [ 948.609341] env[63021]: _type = "Task" [ 948.609341] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.617440] env[63021]: DEBUG oslo_vmware.api [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]520fe007-e5f0-67ab-5b54-dbc133537304, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.654746] env[63021]: DEBUG nova.compute.manager [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 948.658192] env[63021]: DEBUG nova.network.neutron [req-d01e3e11-d605-45af-b3ff-825fcd03ad2b req-ae0cfa25-5b37-4a8f-b9af-815f1d3a4435 service nova] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Updated VIF entry in instance network info cache for port bb825b85-647f-45b0-b351-ec7921e3ae5e. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 948.658608] env[63021]: DEBUG nova.network.neutron [req-d01e3e11-d605-45af-b3ff-825fcd03ad2b req-ae0cfa25-5b37-4a8f-b9af-815f1d3a4435 service nova] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Updating instance_info_cache with network_info: [{"id": "bb825b85-647f-45b0-b351-ec7921e3ae5e", "address": "fa:16:3e:fd:c3:4f", "network": {"id": "265e2e2e-6ec6-4d4c-8b4c-0a2e8baa6bdb", "bridge": "br-int", "label": "tempest-ImagesTestJSON-242691215-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7331caed42034dfc81c79cfabbe0829e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bfae3ef8-cae7-455d-8632-ba93e1671625", "external-id": "cl2-zone-841", "segmentation_id": 841, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb825b85-64", "ovs_interfaceid": "bb825b85-647f-45b0-b351-ec7921e3ae5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.779591] env[63021]: INFO nova.compute.manager [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Took 16.59 seconds to build instance. [ 948.874176] env[63021]: DEBUG nova.scheduler.client.report [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 948.900398] env[63021]: INFO nova.compute.manager [None req-db9db38f-ae08-4a7b-8cb1-5d2701733145 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Took 0.56 seconds to detach 1 volumes for instance. [ 948.902751] env[63021]: DEBUG nova.compute.manager [None req-db9db38f-ae08-4a7b-8cb1-5d2701733145 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Deleting volume: e424adc5-a8b0-4409-b3e0-f3cdbc8fb712 {{(pid=63021) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 949.069993] env[63021]: DEBUG oslo_vmware.api [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294044, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.121655] env[63021]: DEBUG oslo_vmware.api [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]520fe007-e5f0-67ab-5b54-dbc133537304, 'name': SearchDatastore_Task, 'duration_secs': 0.008113} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.122748] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2c571a8-eaf4-4537-af0e-d41eb638a840 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.129436] env[63021]: DEBUG oslo_vmware.api [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 949.129436] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52f9e25e-afc8-6112-ccd7-f745abf7d296" [ 949.129436] env[63021]: _type = "Task" [ 949.129436] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.138649] env[63021]: DEBUG oslo_vmware.api [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52f9e25e-afc8-6112-ccd7-f745abf7d296, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.166339] env[63021]: DEBUG oslo_concurrency.lockutils [req-d01e3e11-d605-45af-b3ff-825fcd03ad2b req-ae0cfa25-5b37-4a8f-b9af-815f1d3a4435 service nova] Releasing lock "refresh_cache-919887c8-171d-48fe-8c6b-2abf5f55d6ab" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.166472] env[63021]: DEBUG nova.compute.manager [req-d01e3e11-d605-45af-b3ff-825fcd03ad2b req-ae0cfa25-5b37-4a8f-b9af-815f1d3a4435 service nova] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Received event network-vif-deleted-3dc5a6b8-23fe-42fa-88c1-6854a326150d {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 949.166701] env[63021]: INFO nova.compute.manager [req-d01e3e11-d605-45af-b3ff-825fcd03ad2b req-ae0cfa25-5b37-4a8f-b9af-815f1d3a4435 service nova] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Neutron deleted interface 3dc5a6b8-23fe-42fa-88c1-6854a326150d; detaching it from the instance and deleting it from the info cache [ 949.166891] env[63021]: DEBUG nova.network.neutron [req-d01e3e11-d605-45af-b3ff-825fcd03ad2b req-ae0cfa25-5b37-4a8f-b9af-815f1d3a4435 service nova] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.187055] env[63021]: DEBUG oslo_concurrency.lockutils [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.282750] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4ef36a0d-3ec0-4692-8c55-78a053e810af tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Lock "a52e4cb5-d0ea-4698-9955-753626762a02" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.107s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.375154] env[63021]: DEBUG nova.compute.manager [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Stashing vm_state: active {{(pid=63021) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 949.389021] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.379s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.389021] env[63021]: DEBUG nova.compute.manager [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 949.390130] env[63021]: DEBUG oslo_concurrency.lockutils [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.508s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.390130] env[63021]: DEBUG nova.objects.instance [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lazy-loading 'pci_requests' on Instance uuid 45c9e2be-eb19-4b83-b280-c9eeaddfccba {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 949.459871] env[63021]: DEBUG oslo_concurrency.lockutils [None req-db9db38f-ae08-4a7b-8cb1-5d2701733145 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.566888] env[63021]: DEBUG oslo_vmware.api [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294044, 'name': PowerOnVM_Task, 'duration_secs': 0.666587} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.567556] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 949.570896] env[63021]: DEBUG nova.compute.manager [None req-600daf06-1406-4d94-b69a-ea9e32f1429f tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 949.571760] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e05500c5-994d-4c3f-95d8-24c12a588f4e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.639842] env[63021]: DEBUG oslo_vmware.api [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52f9e25e-afc8-6112-ccd7-f745abf7d296, 'name': SearchDatastore_Task, 'duration_secs': 0.010484} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.639842] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.640153] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 919887c8-171d-48fe-8c6b-2abf5f55d6ab/919887c8-171d-48fe-8c6b-2abf5f55d6ab.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 949.640538] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7a04e4f8-f7a1-46c3-be23-aced508cf2f1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.647837] env[63021]: DEBUG oslo_vmware.api [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 949.647837] env[63021]: value = "task-1294046" [ 949.647837] env[63021]: _type = "Task" [ 949.647837] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.661780] env[63021]: DEBUG oslo_vmware.api [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294046, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.672292] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5b56a678-d192-464a-a4d7-676fd2b24470 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.683180] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7175bdc6-8f5e-44af-a03a-e4fb1d71dbdd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.715498] env[63021]: DEBUG nova.compute.manager [req-d01e3e11-d605-45af-b3ff-825fcd03ad2b req-ae0cfa25-5b37-4a8f-b9af-815f1d3a4435 service nova] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Detach interface failed, port_id=3dc5a6b8-23fe-42fa-88c1-6854a326150d, reason: Instance 80b857c4-d9dd-4483-970f-a0f5f4bcf173 could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 949.775047] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "cee7a86d-91e2-4b57-b6cd-29056da6a414" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.775047] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "cee7a86d-91e2-4b57-b6cd-29056da6a414" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.894375] env[63021]: DEBUG nova.compute.utils [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 949.900017] env[63021]: DEBUG nova.objects.instance [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lazy-loading 'numa_topology' on Instance uuid 45c9e2be-eb19-4b83-b280-c9eeaddfccba {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 949.901339] env[63021]: DEBUG nova.compute.manager [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 949.901513] env[63021]: DEBUG nova.network.neutron [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 949.908793] env[63021]: DEBUG oslo_concurrency.lockutils [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.958780] env[63021]: DEBUG nova.policy [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6ea19453dbdb4251a93dde2576aeff64', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '58693d1e991249f28914a39761cc2627', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 950.158629] env[63021]: DEBUG oslo_vmware.api [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294046, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.487829} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.159445] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 919887c8-171d-48fe-8c6b-2abf5f55d6ab/919887c8-171d-48fe-8c6b-2abf5f55d6ab.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 950.159445] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 950.159445] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d05d4d3c-6359-4337-83a9-524ee737d6be {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.165521] env[63021]: DEBUG oslo_vmware.api [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 950.165521] env[63021]: value = "task-1294047" [ 950.165521] env[63021]: _type = "Task" [ 950.165521] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.176015] env[63021]: DEBUG oslo_vmware.api [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294047, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.278085] env[63021]: DEBUG nova.compute.manager [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 950.305027] env[63021]: DEBUG nova.network.neutron [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Successfully created port: 3752044a-9145-469d-8edb-6255d7d0f928 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 950.402593] env[63021]: DEBUG nova.compute.manager [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 950.406074] env[63021]: INFO nova.compute.claims [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 950.449586] env[63021]: INFO nova.compute.manager [None req-8e4aceda-aa4c-4e0b-ba00-3fdfa83d7d67 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Unrescuing [ 950.449876] env[63021]: DEBUG oslo_concurrency.lockutils [None req-8e4aceda-aa4c-4e0b-ba00-3fdfa83d7d67 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "refresh_cache-ae103118-bddf-46fe-90b9-98b60952ebba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.450221] env[63021]: DEBUG oslo_concurrency.lockutils [None req-8e4aceda-aa4c-4e0b-ba00-3fdfa83d7d67 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquired lock "refresh_cache-ae103118-bddf-46fe-90b9-98b60952ebba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.450483] env[63021]: DEBUG nova.network.neutron [None req-8e4aceda-aa4c-4e0b-ba00-3fdfa83d7d67 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 950.517620] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Acquiring lock "a52e4cb5-d0ea-4698-9955-753626762a02" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.518234] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Lock "a52e4cb5-d0ea-4698-9955-753626762a02" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.518234] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Acquiring lock "a52e4cb5-d0ea-4698-9955-753626762a02-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.518490] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Lock "a52e4cb5-d0ea-4698-9955-753626762a02-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.518716] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Lock "a52e4cb5-d0ea-4698-9955-753626762a02-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.522087] env[63021]: INFO nova.compute.manager [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Terminating instance [ 950.526130] env[63021]: DEBUG nova.compute.manager [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 950.526381] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 950.527257] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19985565-15d0-4751-b9d7-46f27e9116a0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.538694] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 950.538694] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1ffd9590-88d4-46f0-ad0b-23c19b12ab7d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.544310] env[63021]: DEBUG oslo_vmware.api [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Waiting for the task: (returnval){ [ 950.544310] env[63021]: value = "task-1294048" [ 950.544310] env[63021]: _type = "Task" [ 950.544310] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.552711] env[63021]: DEBUG oslo_vmware.api [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Task: {'id': task-1294048, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.676590] env[63021]: DEBUG oslo_vmware.api [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294047, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063277} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.677611] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 950.681026] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5b02e1b-6b9a-4e26-9463-d4e0f8b9a585 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.705722] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] 919887c8-171d-48fe-8c6b-2abf5f55d6ab/919887c8-171d-48fe-8c6b-2abf5f55d6ab.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 950.706094] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-72923bbf-dce9-40a9-8206-388d5e100914 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.724537] env[63021]: DEBUG oslo_vmware.api [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 950.724537] env[63021]: value = "task-1294049" [ 950.724537] env[63021]: _type = "Task" [ 950.724537] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.733221] env[63021]: DEBUG oslo_vmware.api [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294049, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.798708] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.056655] env[63021]: DEBUG oslo_vmware.api [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Task: {'id': task-1294048, 'name': PowerOffVM_Task, 'duration_secs': 0.38653} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.057022] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 951.057153] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 951.057567] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ccfa8f9a-9e04-4c84-aa7f-ceeb1a97cd56 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.129260] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 951.129532] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Deleting contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 951.129715] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Deleting the datastore file [datastore2] a52e4cb5-d0ea-4698-9955-753626762a02 {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 951.129965] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7fe8fbab-aeb1-4b19-b8b9-428915fe59c5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.138597] env[63021]: DEBUG oslo_vmware.api [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Waiting for the task: (returnval){ [ 951.138597] env[63021]: value = "task-1294051" [ 951.138597] env[63021]: _type = "Task" [ 951.138597] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.149073] env[63021]: DEBUG oslo_vmware.api [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Task: {'id': task-1294051, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.236376] env[63021]: DEBUG oslo_vmware.api [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294049, 'name': ReconfigVM_Task, 'duration_secs': 0.283799} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.236376] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Reconfigured VM instance instance-00000057 to attach disk [datastore2] 919887c8-171d-48fe-8c6b-2abf5f55d6ab/919887c8-171d-48fe-8c6b-2abf5f55d6ab.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 951.237499] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f091d34e-e9c2-4130-82a7-308ab0f9a503 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.246961] env[63021]: DEBUG oslo_vmware.api [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 951.246961] env[63021]: value = "task-1294052" [ 951.246961] env[63021]: _type = "Task" [ 951.246961] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.258060] env[63021]: DEBUG oslo_vmware.api [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294052, 'name': Rename_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.414921] env[63021]: DEBUG nova.compute.manager [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 951.455281] env[63021]: DEBUG nova.virt.hardware [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 951.455502] env[63021]: DEBUG nova.virt.hardware [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 951.455712] env[63021]: DEBUG nova.virt.hardware [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 951.455916] env[63021]: DEBUG nova.virt.hardware [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 951.459873] env[63021]: DEBUG nova.virt.hardware [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 951.460218] env[63021]: DEBUG nova.virt.hardware [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 951.460476] env[63021]: DEBUG nova.virt.hardware [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 951.460826] env[63021]: DEBUG nova.virt.hardware [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 951.461093] env[63021]: DEBUG nova.virt.hardware [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 951.461351] env[63021]: DEBUG nova.virt.hardware [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 951.461610] env[63021]: DEBUG nova.virt.hardware [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 951.462740] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6daf32fd-37e9-4ce9-8a85-bdcb1962ba39 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.478827] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae504770-ee52-4c65-9bcb-60592e1cac0f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.538788] env[63021]: DEBUG nova.network.neutron [None req-8e4aceda-aa4c-4e0b-ba00-3fdfa83d7d67 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Updating instance_info_cache with network_info: [{"id": "f99a201d-df02-499f-85d9-2566ac1dd82b", "address": "fa:16:3e:a0:fd:ff", "network": {"id": "3aaeed35-f452-4144-a259-8f440f8a8bf4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-229941363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b672b1d7d7ad454ebfa3e2fd2421c754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf99a201d-df", "ovs_interfaceid": "f99a201d-df02-499f-85d9-2566ac1dd82b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.648467] env[63021]: DEBUG oslo_vmware.api [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Task: {'id': task-1294051, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.180037} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.648572] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 951.648757] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Deleted contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 951.648938] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 951.649125] env[63021]: INFO nova.compute.manager [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Took 1.12 seconds to destroy the instance on the hypervisor. [ 951.649366] env[63021]: DEBUG oslo.service.loopingcall [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 951.650483] env[63021]: DEBUG nova.compute.manager [-] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 951.650574] env[63021]: DEBUG nova.network.neutron [-] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 951.652882] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-582167d4-694f-4b1f-a765-733d592f9aab {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.660466] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61372c1f-3649-4b22-9476-bf94539e9752 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.693536] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6efd012d-6f90-44b5-aa8d-7cad473d796b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.705984] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba001756-4cae-48e3-983a-3cfbde40a366 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.720575] env[63021]: DEBUG nova.compute.provider_tree [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 951.757024] env[63021]: DEBUG oslo_vmware.api [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294052, 'name': Rename_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.837398] env[63021]: DEBUG nova.network.neutron [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Successfully updated port: 3752044a-9145-469d-8edb-6255d7d0f928 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 951.907676] env[63021]: DEBUG nova.compute.manager [req-1b44ed53-52a8-4eb2-891f-caf869b93242 req-4092a039-3e9d-480c-95d9-a743d1beab2b service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Received event network-vif-plugged-3752044a-9145-469d-8edb-6255d7d0f928 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 951.907676] env[63021]: DEBUG oslo_concurrency.lockutils [req-1b44ed53-52a8-4eb2-891f-caf869b93242 req-4092a039-3e9d-480c-95d9-a743d1beab2b service nova] Acquiring lock "6155c098-e4a6-47e6-b343-4a77ca90eb2e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.907676] env[63021]: DEBUG oslo_concurrency.lockutils [req-1b44ed53-52a8-4eb2-891f-caf869b93242 req-4092a039-3e9d-480c-95d9-a743d1beab2b service nova] Lock "6155c098-e4a6-47e6-b343-4a77ca90eb2e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.907796] env[63021]: DEBUG oslo_concurrency.lockutils [req-1b44ed53-52a8-4eb2-891f-caf869b93242 req-4092a039-3e9d-480c-95d9-a743d1beab2b service nova] Lock "6155c098-e4a6-47e6-b343-4a77ca90eb2e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.907920] env[63021]: DEBUG nova.compute.manager [req-1b44ed53-52a8-4eb2-891f-caf869b93242 req-4092a039-3e9d-480c-95d9-a743d1beab2b service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] No waiting events found dispatching network-vif-plugged-3752044a-9145-469d-8edb-6255d7d0f928 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 951.908211] env[63021]: WARNING nova.compute.manager [req-1b44ed53-52a8-4eb2-891f-caf869b93242 req-4092a039-3e9d-480c-95d9-a743d1beab2b service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Received unexpected event network-vif-plugged-3752044a-9145-469d-8edb-6255d7d0f928 for instance with vm_state building and task_state spawning. [ 952.042909] env[63021]: DEBUG oslo_concurrency.lockutils [None req-8e4aceda-aa4c-4e0b-ba00-3fdfa83d7d67 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Releasing lock "refresh_cache-ae103118-bddf-46fe-90b9-98b60952ebba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.043072] env[63021]: DEBUG nova.objects.instance [None req-8e4aceda-aa4c-4e0b-ba00-3fdfa83d7d67 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lazy-loading 'flavor' on Instance uuid ae103118-bddf-46fe-90b9-98b60952ebba {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 952.086068] env[63021]: DEBUG nova.compute.manager [req-f4c1b7ac-aed2-4295-a787-212a3273271d req-05b2e704-6e5b-481b-b010-8aa32749dd5d service nova] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Received event network-vif-deleted-45d08a72-44cf-4326-b881-7b60e5d8d962 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 952.086276] env[63021]: INFO nova.compute.manager [req-f4c1b7ac-aed2-4295-a787-212a3273271d req-05b2e704-6e5b-481b-b010-8aa32749dd5d service nova] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Neutron deleted interface 45d08a72-44cf-4326-b881-7b60e5d8d962; detaching it from the instance and deleting it from the info cache [ 952.086458] env[63021]: DEBUG nova.network.neutron [req-f4c1b7ac-aed2-4295-a787-212a3273271d req-05b2e704-6e5b-481b-b010-8aa32749dd5d service nova] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.101318] env[63021]: DEBUG oslo_concurrency.lockutils [None req-a933ca2f-0f71-4980-9b4b-45a994f4c93f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquiring lock "df856146-61ba-4de7-a5fd-d75263927163" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.101575] env[63021]: DEBUG oslo_concurrency.lockutils [None req-a933ca2f-0f71-4980-9b4b-45a994f4c93f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "df856146-61ba-4de7-a5fd-d75263927163" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.225068] env[63021]: DEBUG nova.scheduler.client.report [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 952.258102] env[63021]: DEBUG oslo_vmware.api [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294052, 'name': Rename_Task, 'duration_secs': 0.916716} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.258414] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 952.258661] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0180cb09-7592-4e86-b994-db349f5d2316 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.264347] env[63021]: DEBUG oslo_vmware.api [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 952.264347] env[63021]: value = "task-1294053" [ 952.264347] env[63021]: _type = "Task" [ 952.264347] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.274758] env[63021]: DEBUG oslo_vmware.api [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294053, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.340600] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "refresh_cache-6155c098-e4a6-47e6-b343-4a77ca90eb2e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.340828] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquired lock "refresh_cache-6155c098-e4a6-47e6-b343-4a77ca90eb2e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.341050] env[63021]: DEBUG nova.network.neutron [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 952.550785] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f027953e-f8f1-48c1-ae1e-fdadda88a267 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.572114] env[63021]: DEBUG nova.network.neutron [-] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.573433] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e4aceda-aa4c-4e0b-ba00-3fdfa83d7d67 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 952.573952] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e408837e-5b6b-496b-a7d2-ef71c2a7459c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.580853] env[63021]: DEBUG oslo_vmware.api [None req-8e4aceda-aa4c-4e0b-ba00-3fdfa83d7d67 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 952.580853] env[63021]: value = "task-1294054" [ 952.580853] env[63021]: _type = "Task" [ 952.580853] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.595024] env[63021]: DEBUG oslo_vmware.api [None req-8e4aceda-aa4c-4e0b-ba00-3fdfa83d7d67 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294054, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.595319] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ec5898d0-cc86-4bae-a6b3-bac980f3aaba {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.604695] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9465ff03-02ba-4066-bf6d-d2cd644a03e0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.617569] env[63021]: INFO nova.compute.manager [None req-a933ca2f-0f71-4980-9b4b-45a994f4c93f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Detaching volume 1d748e34-0bec-4d79-9a3a-f42d78520ff7 [ 952.640404] env[63021]: DEBUG nova.compute.manager [req-f4c1b7ac-aed2-4295-a787-212a3273271d req-05b2e704-6e5b-481b-b010-8aa32749dd5d service nova] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Detach interface failed, port_id=45d08a72-44cf-4326-b881-7b60e5d8d962, reason: Instance a52e4cb5-d0ea-4698-9955-753626762a02 could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 952.664199] env[63021]: INFO nova.virt.block_device [None req-a933ca2f-0f71-4980-9b4b-45a994f4c93f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Attempting to driver detach volume 1d748e34-0bec-4d79-9a3a-f42d78520ff7 from mountpoint /dev/sdb [ 952.664470] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-a933ca2f-0f71-4980-9b4b-45a994f4c93f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Volume detach. Driver type: vmdk {{(pid=63021) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 952.664665] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-a933ca2f-0f71-4980-9b4b-45a994f4c93f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277540', 'volume_id': '1d748e34-0bec-4d79-9a3a-f42d78520ff7', 'name': 'volume-1d748e34-0bec-4d79-9a3a-f42d78520ff7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'df856146-61ba-4de7-a5fd-d75263927163', 'attached_at': '', 'detached_at': '', 'volume_id': '1d748e34-0bec-4d79-9a3a-f42d78520ff7', 'serial': '1d748e34-0bec-4d79-9a3a-f42d78520ff7'} {{(pid=63021) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 952.665561] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2b4b6af-ae56-419a-b46a-b7b8ac32bb35 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.689263] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73c87656-c5bb-40a6-9616-201790dba353 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.697361] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fb4c966-e563-42b7-9194-7e25c875828e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.719720] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e336cad6-27c1-42b3-8c58-fe7c20f422db {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.736022] env[63021]: DEBUG oslo_concurrency.lockutils [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.346s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.738290] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-a933ca2f-0f71-4980-9b4b-45a994f4c93f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] The volume has not been displaced from its original location: [datastore2] volume-1d748e34-0bec-4d79-9a3a-f42d78520ff7/volume-1d748e34-0bec-4d79-9a3a-f42d78520ff7.vmdk. No consolidation needed. {{(pid=63021) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 952.743952] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-a933ca2f-0f71-4980-9b4b-45a994f4c93f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Reconfiguring VM instance instance-00000043 to detach disk 2001 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 952.744386] env[63021]: DEBUG oslo_concurrency.lockutils [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.141s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.746852] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9f4d4afb-2045-453c-9bb5-87b29a2fb98d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.766624] env[63021]: DEBUG oslo_vmware.api [None req-a933ca2f-0f71-4980-9b4b-45a994f4c93f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 952.766624] env[63021]: value = "task-1294055" [ 952.766624] env[63021]: _type = "Task" [ 952.766624] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.776983] env[63021]: DEBUG oslo_vmware.api [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294053, 'name': PowerOnVM_Task, 'duration_secs': 0.458592} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.780113] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 952.780337] env[63021]: INFO nova.compute.manager [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Took 7.31 seconds to spawn the instance on the hypervisor. [ 952.780513] env[63021]: DEBUG nova.compute.manager [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 952.780814] env[63021]: DEBUG oslo_vmware.api [None req-a933ca2f-0f71-4980-9b4b-45a994f4c93f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294055, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.783812] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24e73e43-5661-4ec7-83b6-a14de110fa9c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.829044] env[63021]: INFO nova.network.neutron [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Updating port 43e48c53-d06d-41ff-97dd-a7747d5551d9 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 952.850535] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Acquiring lock "ed26dff1-fed0-4baf-ad41-d14850254aec" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.850857] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Lock "ed26dff1-fed0-4baf-ad41-d14850254aec" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.851219] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Acquiring lock "ed26dff1-fed0-4baf-ad41-d14850254aec-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.851479] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Lock "ed26dff1-fed0-4baf-ad41-d14850254aec-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.851720] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Lock "ed26dff1-fed0-4baf-ad41-d14850254aec-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.858177] env[63021]: INFO nova.compute.manager [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Terminating instance [ 952.859840] env[63021]: DEBUG nova.compute.manager [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 952.860127] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 952.861024] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27f28215-7b72-4030-8fab-030dfb137950 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.868923] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 952.871681] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e0279f07-f824-451d-a236-a48cac6575fb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.876200] env[63021]: DEBUG nova.network.neutron [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 952.879651] env[63021]: DEBUG oslo_vmware.api [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Waiting for the task: (returnval){ [ 952.879651] env[63021]: value = "task-1294056" [ 952.879651] env[63021]: _type = "Task" [ 952.879651] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.893022] env[63021]: DEBUG oslo_vmware.api [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Task: {'id': task-1294056, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.990619] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a81e75c6-3f59-43f0-8b72-6fa151cf0fb3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.998587] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d55b6456-bc41-4caf-85d6-9ffa2c9d0b7b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.032806] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c50dff6-e2f3-4bbb-9718-1a5c75e4b311 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.041190] env[63021]: DEBUG nova.network.neutron [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Updating instance_info_cache with network_info: [{"id": "3752044a-9145-469d-8edb-6255d7d0f928", "address": "fa:16:3e:fe:57:1f", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3752044a-91", "ovs_interfaceid": "3752044a-9145-469d-8edb-6255d7d0f928", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.042922] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1260e28b-3191-4e9d-b656-80dce6c5168b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.059612] env[63021]: DEBUG nova.compute.provider_tree [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 953.076368] env[63021]: INFO nova.compute.manager [-] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Took 1.42 seconds to deallocate network for instance. [ 953.089686] env[63021]: DEBUG oslo_vmware.api [None req-8e4aceda-aa4c-4e0b-ba00-3fdfa83d7d67 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294054, 'name': PowerOffVM_Task, 'duration_secs': 0.3352} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.089961] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e4aceda-aa4c-4e0b-ba00-3fdfa83d7d67 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 953.096735] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e4aceda-aa4c-4e0b-ba00-3fdfa83d7d67 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Reconfiguring VM instance instance-00000052 to detach disk 2001 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 953.097064] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a018cf33-e70a-430a-b5d7-b265ee966622 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.114740] env[63021]: DEBUG oslo_vmware.api [None req-8e4aceda-aa4c-4e0b-ba00-3fdfa83d7d67 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 953.114740] env[63021]: value = "task-1294057" [ 953.114740] env[63021]: _type = "Task" [ 953.114740] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.123085] env[63021]: DEBUG oslo_vmware.api [None req-8e4aceda-aa4c-4e0b-ba00-3fdfa83d7d67 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294057, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.279779] env[63021]: DEBUG oslo_vmware.api [None req-a933ca2f-0f71-4980-9b4b-45a994f4c93f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294055, 'name': ReconfigVM_Task, 'duration_secs': 0.247646} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.280117] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-a933ca2f-0f71-4980-9b4b-45a994f4c93f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Reconfigured VM instance instance-00000043 to detach disk 2001 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 953.284864] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-599355d2-7d0f-4b95-ac22-37f704dff698 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.301774] env[63021]: INFO nova.compute.manager [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Took 14.62 seconds to build instance. [ 953.303848] env[63021]: DEBUG oslo_vmware.api [None req-a933ca2f-0f71-4980-9b4b-45a994f4c93f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 953.303848] env[63021]: value = "task-1294058" [ 953.303848] env[63021]: _type = "Task" [ 953.303848] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.312976] env[63021]: DEBUG oslo_vmware.api [None req-a933ca2f-0f71-4980-9b4b-45a994f4c93f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294058, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.395796] env[63021]: DEBUG oslo_vmware.api [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Task: {'id': task-1294056, 'name': PowerOffVM_Task, 'duration_secs': 0.291218} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.396258] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 953.396555] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 953.396927] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-63c924d3-1a5a-4c84-8d73-ff846c795fb2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.466766] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 953.467033] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 953.467233] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Deleting the datastore file [datastore1] ed26dff1-fed0-4baf-ad41-d14850254aec {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 953.467504] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4c319633-07a2-49a9-a042-0fccf4f275fa {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.474069] env[63021]: DEBUG oslo_vmware.api [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Waiting for the task: (returnval){ [ 953.474069] env[63021]: value = "task-1294060" [ 953.474069] env[63021]: _type = "Task" [ 953.474069] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.482904] env[63021]: DEBUG oslo_vmware.api [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Task: {'id': task-1294060, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.547653] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Releasing lock "refresh_cache-6155c098-e4a6-47e6-b343-4a77ca90eb2e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.548036] env[63021]: DEBUG nova.compute.manager [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Instance network_info: |[{"id": "3752044a-9145-469d-8edb-6255d7d0f928", "address": "fa:16:3e:fe:57:1f", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3752044a-91", "ovs_interfaceid": "3752044a-9145-469d-8edb-6255d7d0f928", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 953.548507] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fe:57:1f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8059554c-499f-44b4-be06-29f80ec36b34', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3752044a-9145-469d-8edb-6255d7d0f928', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 953.556147] env[63021]: DEBUG oslo.service.loopingcall [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 953.556393] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 953.556626] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9f57e7cc-3977-4023-b9b8-1de9a1cd551b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.571651] env[63021]: DEBUG nova.scheduler.client.report [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 953.579847] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 953.579847] env[63021]: value = "task-1294061" [ 953.579847] env[63021]: _type = "Task" [ 953.579847] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.580805] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.589230] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294061, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.625013] env[63021]: DEBUG oslo_vmware.api [None req-8e4aceda-aa4c-4e0b-ba00-3fdfa83d7d67 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294057, 'name': ReconfigVM_Task, 'duration_secs': 0.427949} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.625410] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e4aceda-aa4c-4e0b-ba00-3fdfa83d7d67 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Reconfigured VM instance instance-00000052 to detach disk 2001 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 953.625641] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e4aceda-aa4c-4e0b-ba00-3fdfa83d7d67 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 953.625974] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ec334fc9-0346-46cc-b2f6-a46191c51ef5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.631688] env[63021]: DEBUG oslo_vmware.api [None req-8e4aceda-aa4c-4e0b-ba00-3fdfa83d7d67 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 953.631688] env[63021]: value = "task-1294062" [ 953.631688] env[63021]: _type = "Task" [ 953.631688] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.639667] env[63021]: DEBUG oslo_vmware.api [None req-8e4aceda-aa4c-4e0b-ba00-3fdfa83d7d67 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294062, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.809018] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4958f216-b985-4ce2-8789-d9e366cc0d77 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "919887c8-171d-48fe-8c6b-2abf5f55d6ab" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.134s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.814526] env[63021]: DEBUG oslo_vmware.api [None req-a933ca2f-0f71-4980-9b4b-45a994f4c93f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294058, 'name': ReconfigVM_Task, 'duration_secs': 0.150698} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.814863] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-a933ca2f-0f71-4980-9b4b-45a994f4c93f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277540', 'volume_id': '1d748e34-0bec-4d79-9a3a-f42d78520ff7', 'name': 'volume-1d748e34-0bec-4d79-9a3a-f42d78520ff7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'df856146-61ba-4de7-a5fd-d75263927163', 'attached_at': '', 'detached_at': '', 'volume_id': '1d748e34-0bec-4d79-9a3a-f42d78520ff7', 'serial': '1d748e34-0bec-4d79-9a3a-f42d78520ff7'} {{(pid=63021) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 953.984380] env[63021]: DEBUG oslo_vmware.api [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Task: {'id': task-1294060, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.269298} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.984661] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 953.984894] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 953.985102] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 953.985282] env[63021]: INFO nova.compute.manager [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Took 1.13 seconds to destroy the instance on the hypervisor. [ 953.985528] env[63021]: DEBUG oslo.service.loopingcall [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 953.985726] env[63021]: DEBUG nova.compute.manager [-] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 953.985824] env[63021]: DEBUG nova.network.neutron [-] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 954.003949] env[63021]: DEBUG nova.compute.manager [req-108f0c0d-2807-4e79-afc4-03d66e154967 req-ea673beb-f7fa-40da-9f63-fbe19a08a3a6 service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Received event network-changed-3752044a-9145-469d-8edb-6255d7d0f928 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 954.004291] env[63021]: DEBUG nova.compute.manager [req-108f0c0d-2807-4e79-afc4-03d66e154967 req-ea673beb-f7fa-40da-9f63-fbe19a08a3a6 service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Refreshing instance network info cache due to event network-changed-3752044a-9145-469d-8edb-6255d7d0f928. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 954.004578] env[63021]: DEBUG oslo_concurrency.lockutils [req-108f0c0d-2807-4e79-afc4-03d66e154967 req-ea673beb-f7fa-40da-9f63-fbe19a08a3a6 service nova] Acquiring lock "refresh_cache-6155c098-e4a6-47e6-b343-4a77ca90eb2e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.004784] env[63021]: DEBUG oslo_concurrency.lockutils [req-108f0c0d-2807-4e79-afc4-03d66e154967 req-ea673beb-f7fa-40da-9f63-fbe19a08a3a6 service nova] Acquired lock "refresh_cache-6155c098-e4a6-47e6-b343-4a77ca90eb2e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.004976] env[63021]: DEBUG nova.network.neutron [req-108f0c0d-2807-4e79-afc4-03d66e154967 req-ea673beb-f7fa-40da-9f63-fbe19a08a3a6 service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Refreshing network info cache for port 3752044a-9145-469d-8edb-6255d7d0f928 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 954.078059] env[63021]: DEBUG oslo_concurrency.lockutils [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.332s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.078059] env[63021]: INFO nova.compute.manager [None req-08ab2a11-8bf5-44bb-9a40-f9ed84709ebb tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Successfully reverted task state from rebuilding on failure for instance. [ 954.087609] env[63021]: DEBUG oslo_concurrency.lockutils [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.901s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.090179] env[63021]: INFO nova.compute.claims [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 954.104702] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294061, 'name': CreateVM_Task, 'duration_secs': 0.344493} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.104702] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 954.105793] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.105990] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.106329] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 954.106624] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-171f0b18-8d28-45fd-a13d-8ec3ba027ae8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.116011] env[63021]: DEBUG oslo_vmware.api [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 954.116011] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52819364-d374-c98d-45e0-e1ed29f916e3" [ 954.116011] env[63021]: _type = "Task" [ 954.116011] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.124870] env[63021]: DEBUG oslo_vmware.api [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52819364-d374-c98d-45e0-e1ed29f916e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.146464] env[63021]: DEBUG oslo_vmware.api [None req-8e4aceda-aa4c-4e0b-ba00-3fdfa83d7d67 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294062, 'name': PowerOnVM_Task, 'duration_secs': 0.432709} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.146821] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e4aceda-aa4c-4e0b-ba00-3fdfa83d7d67 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 954.147451] env[63021]: DEBUG nova.compute.manager [None req-8e4aceda-aa4c-4e0b-ba00-3fdfa83d7d67 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 954.148096] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6de4a4f4-b6cb-4412-827d-5a61f8525f55 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.364534] env[63021]: DEBUG nova.objects.instance [None req-a933ca2f-0f71-4980-9b4b-45a994f4c93f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lazy-loading 'flavor' on Instance uuid df856146-61ba-4de7-a5fd-d75263927163 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 954.487130] env[63021]: DEBUG nova.compute.manager [req-0ca387d8-611d-4ad9-93f5-41c6bc30a427 req-80cb56e2-a53e-47e5-b9ff-1c7eae49cbde service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Received event network-vif-plugged-43e48c53-d06d-41ff-97dd-a7747d5551d9 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 954.487381] env[63021]: DEBUG oslo_concurrency.lockutils [req-0ca387d8-611d-4ad9-93f5-41c6bc30a427 req-80cb56e2-a53e-47e5-b9ff-1c7eae49cbde service nova] Acquiring lock "45c9e2be-eb19-4b83-b280-c9eeaddfccba-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.487604] env[63021]: DEBUG oslo_concurrency.lockutils [req-0ca387d8-611d-4ad9-93f5-41c6bc30a427 req-80cb56e2-a53e-47e5-b9ff-1c7eae49cbde service nova] Lock "45c9e2be-eb19-4b83-b280-c9eeaddfccba-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.487763] env[63021]: DEBUG oslo_concurrency.lockutils [req-0ca387d8-611d-4ad9-93f5-41c6bc30a427 req-80cb56e2-a53e-47e5-b9ff-1c7eae49cbde service nova] Lock "45c9e2be-eb19-4b83-b280-c9eeaddfccba-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.488223] env[63021]: DEBUG nova.compute.manager [req-0ca387d8-611d-4ad9-93f5-41c6bc30a427 req-80cb56e2-a53e-47e5-b9ff-1c7eae49cbde service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] No waiting events found dispatching network-vif-plugged-43e48c53-d06d-41ff-97dd-a7747d5551d9 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 954.488447] env[63021]: WARNING nova.compute.manager [req-0ca387d8-611d-4ad9-93f5-41c6bc30a427 req-80cb56e2-a53e-47e5-b9ff-1c7eae49cbde service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Received unexpected event network-vif-plugged-43e48c53-d06d-41ff-97dd-a7747d5551d9 for instance with vm_state shelved_offloaded and task_state spawning. [ 954.635538] env[63021]: DEBUG oslo_vmware.api [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52819364-d374-c98d-45e0-e1ed29f916e3, 'name': SearchDatastore_Task, 'duration_secs': 0.017546} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.637023] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.637023] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 954.637441] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.637628] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.637891] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 954.638589] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-12707c6b-7a16-4e93-8156-35fb1d983beb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.647695] env[63021]: DEBUG oslo_concurrency.lockutils [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Acquiring lock "refresh_cache-45c9e2be-eb19-4b83-b280-c9eeaddfccba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.647917] env[63021]: DEBUG oslo_concurrency.lockutils [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Acquired lock "refresh_cache-45c9e2be-eb19-4b83-b280-c9eeaddfccba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.648207] env[63021]: DEBUG nova.network.neutron [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 954.655173] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 954.655449] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 954.657332] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8cdfb6de-111b-4f8c-a634-b601ad3c7b57 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.671558] env[63021]: DEBUG oslo_vmware.api [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 954.671558] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52e3c0ca-754c-ebea-8d00-f98501957c6b" [ 954.671558] env[63021]: _type = "Task" [ 954.671558] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.687208] env[63021]: DEBUG oslo_vmware.api [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52e3c0ca-754c-ebea-8d00-f98501957c6b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.720125] env[63021]: DEBUG nova.compute.manager [None req-b5ea11b7-7b51-4404-9b05-ba2dae0b310c tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 954.720125] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e009805-b20c-49f7-8489-bdd4332c3444 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.776595] env[63021]: DEBUG nova.network.neutron [req-108f0c0d-2807-4e79-afc4-03d66e154967 req-ea673beb-f7fa-40da-9f63-fbe19a08a3a6 service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Updated VIF entry in instance network info cache for port 3752044a-9145-469d-8edb-6255d7d0f928. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 954.776595] env[63021]: DEBUG nova.network.neutron [req-108f0c0d-2807-4e79-afc4-03d66e154967 req-ea673beb-f7fa-40da-9f63-fbe19a08a3a6 service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Updating instance_info_cache with network_info: [{"id": "3752044a-9145-469d-8edb-6255d7d0f928", "address": "fa:16:3e:fe:57:1f", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3752044a-91", "ovs_interfaceid": "3752044a-9145-469d-8edb-6255d7d0f928", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.950237] env[63021]: DEBUG oslo_concurrency.lockutils [None req-27b0e2ae-955e-4b60-bb43-06e4862f8fa3 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "c328d2f7-3398-4f25-b11c-f464be7af8a1" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.950603] env[63021]: DEBUG oslo_concurrency.lockutils [None req-27b0e2ae-955e-4b60-bb43-06e4862f8fa3 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "c328d2f7-3398-4f25-b11c-f464be7af8a1" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.950885] env[63021]: INFO nova.compute.manager [None req-27b0e2ae-955e-4b60-bb43-06e4862f8fa3 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Rebooting instance [ 955.042362] env[63021]: DEBUG nova.network.neutron [-] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.181972] env[63021]: DEBUG oslo_vmware.api [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52e3c0ca-754c-ebea-8d00-f98501957c6b, 'name': SearchDatastore_Task, 'duration_secs': 0.030538} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.184811] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5979c3cf-f043-499e-892f-b5ff8cc4d883 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.192842] env[63021]: DEBUG oslo_vmware.api [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 955.192842] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5200dc56-1867-eb34-65f0-6e52a8c09480" [ 955.192842] env[63021]: _type = "Task" [ 955.192842] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.203388] env[63021]: DEBUG oslo_vmware.api [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5200dc56-1867-eb34-65f0-6e52a8c09480, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.231703] env[63021]: INFO nova.compute.manager [None req-b5ea11b7-7b51-4404-9b05-ba2dae0b310c tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] instance snapshotting [ 955.238548] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-364808d1-4cc8-4875-a991-19e132997f4d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.259836] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74611d53-23cd-4830-88e2-197025c9748e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.277912] env[63021]: DEBUG oslo_concurrency.lockutils [req-108f0c0d-2807-4e79-afc4-03d66e154967 req-ea673beb-f7fa-40da-9f63-fbe19a08a3a6 service nova] Releasing lock "refresh_cache-6155c098-e4a6-47e6-b343-4a77ca90eb2e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.316033] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4707b161-513e-4e03-bdd9-7b4e2cb61c04 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquiring lock "df856146-61ba-4de7-a5fd-d75263927163" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.332924] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70d7dae1-64bf-4a60-bc75-c942f5aae95e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.340372] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00182019-10b8-44e2-84cc-da8323d4bb3f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.372970] env[63021]: DEBUG oslo_concurrency.lockutils [None req-a933ca2f-0f71-4980-9b4b-45a994f4c93f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "df856146-61ba-4de7-a5fd-d75263927163" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.271s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.376795] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4707b161-513e-4e03-bdd9-7b4e2cb61c04 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "df856146-61ba-4de7-a5fd-d75263927163" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.062s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.377551] env[63021]: DEBUG nova.compute.manager [None req-4707b161-513e-4e03-bdd9-7b4e2cb61c04 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 955.378049] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4755310-49a6-457b-9131-8f137afee3f6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.381877] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bda538b3-b1f8-4f30-81d3-cfc8914e3d15 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.392628] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dc57d56-a6c2-4499-95eb-9a886d6ab52d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.397662] env[63021]: DEBUG nova.compute.manager [None req-4707b161-513e-4e03-bdd9-7b4e2cb61c04 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63021) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 955.397789] env[63021]: DEBUG nova.objects.instance [None req-4707b161-513e-4e03-bdd9-7b4e2cb61c04 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lazy-loading 'flavor' on Instance uuid df856146-61ba-4de7-a5fd-d75263927163 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 955.411063] env[63021]: DEBUG nova.compute.provider_tree [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 955.422751] env[63021]: DEBUG nova.network.neutron [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Updating instance_info_cache with network_info: [{"id": "43e48c53-d06d-41ff-97dd-a7747d5551d9", "address": "fa:16:3e:73:e1:9e", "network": {"id": "36548b96-bd4d-4725-8dab-5cb1cca12c4e", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-672805719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.178", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b4c8cd929ff4068b45ee829b19f9769", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8614db14-cc04-466b-b309-367ab8296cda", "external-id": "nsx-vlan-transportzone-337", "segmentation_id": 337, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43e48c53-d0", "ovs_interfaceid": "43e48c53-d06d-41ff-97dd-a7747d5551d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.477023] env[63021]: DEBUG oslo_concurrency.lockutils [None req-27b0e2ae-955e-4b60-bb43-06e4862f8fa3 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "refresh_cache-c328d2f7-3398-4f25-b11c-f464be7af8a1" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.477023] env[63021]: DEBUG oslo_concurrency.lockutils [None req-27b0e2ae-955e-4b60-bb43-06e4862f8fa3 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquired lock "refresh_cache-c328d2f7-3398-4f25-b11c-f464be7af8a1" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.477023] env[63021]: DEBUG nova.network.neutron [None req-27b0e2ae-955e-4b60-bb43-06e4862f8fa3 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 955.545431] env[63021]: INFO nova.compute.manager [-] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Took 1.56 seconds to deallocate network for instance. [ 955.703413] env[63021]: DEBUG oslo_vmware.api [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5200dc56-1867-eb34-65f0-6e52a8c09480, 'name': SearchDatastore_Task, 'duration_secs': 0.016076} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.703706] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.704289] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 6155c098-e4a6-47e6-b343-4a77ca90eb2e/6155c098-e4a6-47e6-b343-4a77ca90eb2e.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 955.704289] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4ce1b41b-49a9-456f-a3d6-9c7aa8e29627 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.711710] env[63021]: DEBUG oslo_vmware.api [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 955.711710] env[63021]: value = "task-1294063" [ 955.711710] env[63021]: _type = "Task" [ 955.711710] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.721356] env[63021]: DEBUG oslo_vmware.api [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294063, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.772065] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b5ea11b7-7b51-4404-9b05-ba2dae0b310c tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Creating Snapshot of the VM instance {{(pid=63021) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 955.772065] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-0638594f-7033-43c2-b48e-e85238c984f7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.778871] env[63021]: DEBUG oslo_vmware.api [None req-b5ea11b7-7b51-4404-9b05-ba2dae0b310c tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 955.778871] env[63021]: value = "task-1294064" [ 955.778871] env[63021]: _type = "Task" [ 955.778871] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.787353] env[63021]: DEBUG oslo_vmware.api [None req-b5ea11b7-7b51-4404-9b05-ba2dae0b310c tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294064, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.903035] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-4707b161-513e-4e03-bdd9-7b4e2cb61c04 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 955.903341] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-30ec6c85-38c3-4d77-b3b3-84c11ac938ab {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.910361] env[63021]: DEBUG oslo_vmware.api [None req-4707b161-513e-4e03-bdd9-7b4e2cb61c04 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 955.910361] env[63021]: value = "task-1294065" [ 955.910361] env[63021]: _type = "Task" [ 955.910361] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.914284] env[63021]: DEBUG nova.scheduler.client.report [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 955.923521] env[63021]: DEBUG oslo_vmware.api [None req-4707b161-513e-4e03-bdd9-7b4e2cb61c04 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294065, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.926537] env[63021]: DEBUG oslo_concurrency.lockutils [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Releasing lock "refresh_cache-45c9e2be-eb19-4b83-b280-c9eeaddfccba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.960142] env[63021]: DEBUG nova.virt.hardware [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='cf4a784e7201a36a2839be54dc8d2013',container_format='bare',created_at=2024-10-15T17:50:11Z,direct_url=,disk_format='vmdk',id=d9df5b42-2840-4afc-82fa-268aed408575,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-811953221-shelved',owner='4b4c8cd929ff4068b45ee829b19f9769',properties=ImageMetaProps,protected=,size=31667200,status='active',tags=,updated_at=2024-10-15T17:50:32Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 955.960436] env[63021]: DEBUG nova.virt.hardware [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 955.960589] env[63021]: DEBUG nova.virt.hardware [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 955.960775] env[63021]: DEBUG nova.virt.hardware [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 955.960927] env[63021]: DEBUG nova.virt.hardware [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 955.961145] env[63021]: DEBUG nova.virt.hardware [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 955.961381] env[63021]: DEBUG nova.virt.hardware [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 955.961551] env[63021]: DEBUG nova.virt.hardware [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 955.961774] env[63021]: DEBUG nova.virt.hardware [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 955.961967] env[63021]: DEBUG nova.virt.hardware [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 955.962165] env[63021]: DEBUG nova.virt.hardware [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 955.963211] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe8f7368-917d-47c1-ad15-04bce3861283 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.972972] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f80a928-c007-4071-bb69-c9c528af02c6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.991610] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:73:e1:9e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8614db14-cc04-466b-b309-367ab8296cda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '43e48c53-d06d-41ff-97dd-a7747d5551d9', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 955.999997] env[63021]: DEBUG oslo.service.loopingcall [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 956.000702] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 956.001355] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-19f3190c-93d2-4a59-be69-608a35a74ee9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.024134] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 956.024134] env[63021]: value = "task-1294066" [ 956.024134] env[63021]: _type = "Task" [ 956.024134] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.033603] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294066, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.038445] env[63021]: DEBUG nova.compute.manager [req-729388fc-e27c-4b3b-9dee-340f4bb0e317 req-75bdea42-3f9f-43e9-b477-d961f7db35f4 service nova] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Received event network-vif-deleted-2e8a7fc6-825c-45a5-b41c-d28410061aa3 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 956.053549] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.225174] env[63021]: DEBUG oslo_vmware.api [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294063, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.288684] env[63021]: DEBUG oslo_vmware.api [None req-b5ea11b7-7b51-4404-9b05-ba2dae0b310c tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294064, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.421691] env[63021]: DEBUG oslo_concurrency.lockutils [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.334s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.424262] env[63021]: DEBUG nova.compute.manager [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 956.426449] env[63021]: DEBUG oslo_vmware.api [None req-4707b161-513e-4e03-bdd9-7b4e2cb61c04 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294065, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.426942] env[63021]: DEBUG oslo_concurrency.lockutils [None req-db9db38f-ae08-4a7b-8cb1-5d2701733145 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.967s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.427682] env[63021]: DEBUG nova.objects.instance [None req-db9db38f-ae08-4a7b-8cb1-5d2701733145 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Lazy-loading 'resources' on Instance uuid 80b857c4-d9dd-4483-970f-a0f5f4bcf173 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 956.518944] env[63021]: DEBUG nova.compute.manager [req-6ebfb38b-1929-4b3c-882e-aff6428c5deb req-46a060f8-a2e8-44db-86d8-a5d45163591f service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Received event network-changed-43e48c53-d06d-41ff-97dd-a7747d5551d9 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 956.519110] env[63021]: DEBUG nova.compute.manager [req-6ebfb38b-1929-4b3c-882e-aff6428c5deb req-46a060f8-a2e8-44db-86d8-a5d45163591f service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Refreshing instance network info cache due to event network-changed-43e48c53-d06d-41ff-97dd-a7747d5551d9. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 956.519378] env[63021]: DEBUG oslo_concurrency.lockutils [req-6ebfb38b-1929-4b3c-882e-aff6428c5deb req-46a060f8-a2e8-44db-86d8-a5d45163591f service nova] Acquiring lock "refresh_cache-45c9e2be-eb19-4b83-b280-c9eeaddfccba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.519690] env[63021]: DEBUG oslo_concurrency.lockutils [req-6ebfb38b-1929-4b3c-882e-aff6428c5deb req-46a060f8-a2e8-44db-86d8-a5d45163591f service nova] Acquired lock "refresh_cache-45c9e2be-eb19-4b83-b280-c9eeaddfccba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.519764] env[63021]: DEBUG nova.network.neutron [req-6ebfb38b-1929-4b3c-882e-aff6428c5deb req-46a060f8-a2e8-44db-86d8-a5d45163591f service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Refreshing network info cache for port 43e48c53-d06d-41ff-97dd-a7747d5551d9 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 956.537455] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294066, 'name': CreateVM_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.583140] env[63021]: DEBUG nova.network.neutron [None req-27b0e2ae-955e-4b60-bb43-06e4862f8fa3 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Updating instance_info_cache with network_info: [{"id": "eff6f35a-0a13-4e0c-88b0-96b66b240f9b", "address": "fa:16:3e:27:3b:5e", "network": {"id": "9502e364-60e1-47ef-a1f7-f3e3ff45cf69", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-164717784-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5af87713ae6b477ca8a584aed287889d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeff6f35a-0a", "ovs_interfaceid": "eff6f35a-0a13-4e0c-88b0-96b66b240f9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.722337] env[63021]: DEBUG oslo_vmware.api [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294063, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.555332} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.722613] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 6155c098-e4a6-47e6-b343-4a77ca90eb2e/6155c098-e4a6-47e6-b343-4a77ca90eb2e.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 956.722833] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 956.723102] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-994f98eb-4faa-4a93-b352-7ad6880f274f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.733524] env[63021]: DEBUG oslo_vmware.api [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 956.733524] env[63021]: value = "task-1294067" [ 956.733524] env[63021]: _type = "Task" [ 956.733524] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.742842] env[63021]: DEBUG oslo_vmware.api [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294067, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.792945] env[63021]: DEBUG oslo_vmware.api [None req-b5ea11b7-7b51-4404-9b05-ba2dae0b310c tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294064, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.920324] env[63021]: DEBUG oslo_vmware.api [None req-4707b161-513e-4e03-bdd9-7b4e2cb61c04 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294065, 'name': PowerOffVM_Task, 'duration_secs': 0.600308} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.920625] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-4707b161-513e-4e03-bdd9-7b4e2cb61c04 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 956.920808] env[63021]: DEBUG nova.compute.manager [None req-4707b161-513e-4e03-bdd9-7b4e2cb61c04 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 956.921601] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ef321e5-76b9-4b88-82fa-ca549d3647df {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.930357] env[63021]: DEBUG nova.compute.utils [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 956.935109] env[63021]: DEBUG nova.compute.manager [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 956.935109] env[63021]: DEBUG nova.network.neutron [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 956.980246] env[63021]: DEBUG nova.policy [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c1f5022564954345b41649f2a74b5253', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'efc7381b5cb942bf9ef93d6504d783e6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 957.044682] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294066, 'name': CreateVM_Task, 'duration_secs': 0.620259} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.044682] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 957.044682] env[63021]: DEBUG oslo_concurrency.lockutils [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d9df5b42-2840-4afc-82fa-268aed408575" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.044682] env[63021]: DEBUG oslo_concurrency.lockutils [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d9df5b42-2840-4afc-82fa-268aed408575" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.044682] env[63021]: DEBUG oslo_concurrency.lockutils [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d9df5b42-2840-4afc-82fa-268aed408575" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 957.045158] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cebc4704-840c-46b0-9655-d440bd980fcb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.052091] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Waiting for the task: (returnval){ [ 957.052091] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5259fa8f-52ff-1378-a717-e58e769a8b81" [ 957.052091] env[63021]: _type = "Task" [ 957.052091] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.064803] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5259fa8f-52ff-1378-a717-e58e769a8b81, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.086333] env[63021]: DEBUG oslo_concurrency.lockutils [None req-27b0e2ae-955e-4b60-bb43-06e4862f8fa3 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Releasing lock "refresh_cache-c328d2f7-3398-4f25-b11c-f464be7af8a1" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.093079] env[63021]: DEBUG nova.compute.manager [None req-27b0e2ae-955e-4b60-bb43-06e4862f8fa3 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 957.094175] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb64f684-e337-4b85-8bcc-1c07b09b8f25 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.181493] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d36bee55-7e78-4362-befe-651ed970492c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.189209] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ae0daf9-bfe8-4ddc-9f4c-905892b4f8d9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.225505] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-312abc33-d99f-4726-8171-cbc0c4e18430 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.242283] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5837437f-2061-42e5-895f-b8915d3818fe {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.252271] env[63021]: DEBUG oslo_vmware.api [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294067, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.252101} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.261952] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 957.263314] env[63021]: DEBUG nova.compute.provider_tree [None req-db9db38f-ae08-4a7b-8cb1-5d2701733145 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 957.265208] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cddd2ab-7859-4e1a-9a4b-e813786c6f2b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.292666] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] 6155c098-e4a6-47e6-b343-4a77ca90eb2e/6155c098-e4a6-47e6-b343-4a77ca90eb2e.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 957.296742] env[63021]: DEBUG nova.network.neutron [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Successfully created port: a8e61c39-aea7-48e6-9798-cbf508c50425 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 957.301260] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2159e9b9-8266-4e1a-9449-ad10110f25b9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.321962] env[63021]: DEBUG oslo_vmware.api [None req-b5ea11b7-7b51-4404-9b05-ba2dae0b310c tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294064, 'name': CreateSnapshot_Task, 'duration_secs': 1.339729} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.323335] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b5ea11b7-7b51-4404-9b05-ba2dae0b310c tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Created Snapshot of the VM instance {{(pid=63021) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 957.324011] env[63021]: DEBUG oslo_vmware.api [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 957.324011] env[63021]: value = "task-1294068" [ 957.324011] env[63021]: _type = "Task" [ 957.324011] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.326701] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dcbb806-52a5-410b-8c87-0d7c1dd0f88f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.342145] env[63021]: DEBUG oslo_vmware.api [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294068, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.437603] env[63021]: DEBUG nova.compute.manager [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 957.440136] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4707b161-513e-4e03-bdd9-7b4e2cb61c04 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "df856146-61ba-4de7-a5fd-d75263927163" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.063s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.524519] env[63021]: DEBUG nova.network.neutron [req-6ebfb38b-1929-4b3c-882e-aff6428c5deb req-46a060f8-a2e8-44db-86d8-a5d45163591f service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Updated VIF entry in instance network info cache for port 43e48c53-d06d-41ff-97dd-a7747d5551d9. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 957.524900] env[63021]: DEBUG nova.network.neutron [req-6ebfb38b-1929-4b3c-882e-aff6428c5deb req-46a060f8-a2e8-44db-86d8-a5d45163591f service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Updating instance_info_cache with network_info: [{"id": "43e48c53-d06d-41ff-97dd-a7747d5551d9", "address": "fa:16:3e:73:e1:9e", "network": {"id": "36548b96-bd4d-4725-8dab-5cb1cca12c4e", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-672805719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.178", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b4c8cd929ff4068b45ee829b19f9769", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8614db14-cc04-466b-b309-367ab8296cda", "external-id": "nsx-vlan-transportzone-337", "segmentation_id": 337, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43e48c53-d0", "ovs_interfaceid": "43e48c53-d06d-41ff-97dd-a7747d5551d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.564860] env[63021]: DEBUG oslo_concurrency.lockutils [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d9df5b42-2840-4afc-82fa-268aed408575" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.565221] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Processing image d9df5b42-2840-4afc-82fa-268aed408575 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 957.565482] env[63021]: DEBUG oslo_concurrency.lockutils [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d9df5b42-2840-4afc-82fa-268aed408575/d9df5b42-2840-4afc-82fa-268aed408575.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.565636] env[63021]: DEBUG oslo_concurrency.lockutils [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d9df5b42-2840-4afc-82fa-268aed408575/d9df5b42-2840-4afc-82fa-268aed408575.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.565840] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 957.566123] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8787ac4c-cac4-4c23-9e6a-2246bd2377c9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.574168] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 957.574418] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 957.575165] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-775019d1-c8b9-47b2-8a81-b6bef94864a0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.580623] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Waiting for the task: (returnval){ [ 957.580623] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5206c9f1-488a-e525-972a-623ca3fb3d15" [ 957.580623] env[63021]: _type = "Task" [ 957.580623] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.588366] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5206c9f1-488a-e525-972a-623ca3fb3d15, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.769775] env[63021]: DEBUG nova.scheduler.client.report [None req-db9db38f-ae08-4a7b-8cb1-5d2701733145 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 957.837556] env[63021]: DEBUG oslo_vmware.api [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294068, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.849862] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b5ea11b7-7b51-4404-9b05-ba2dae0b310c tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Creating linked-clone VM from snapshot {{(pid=63021) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 957.850596] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-e4b1dde1-f991-4de8-85ca-cdf28ba7aff8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.856798] env[63021]: DEBUG nova.objects.instance [None req-12af6f49-90b6-46d7-b591-53c8cb0191dd tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lazy-loading 'flavor' on Instance uuid df856146-61ba-4de7-a5fd-d75263927163 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 957.859744] env[63021]: DEBUG oslo_vmware.api [None req-b5ea11b7-7b51-4404-9b05-ba2dae0b310c tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 957.859744] env[63021]: value = "task-1294069" [ 957.859744] env[63021]: _type = "Task" [ 957.859744] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.868578] env[63021]: DEBUG oslo_vmware.api [None req-b5ea11b7-7b51-4404-9b05-ba2dae0b310c tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294069, 'name': CloneVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.027784] env[63021]: DEBUG oslo_concurrency.lockutils [req-6ebfb38b-1929-4b3c-882e-aff6428c5deb req-46a060f8-a2e8-44db-86d8-a5d45163591f service nova] Releasing lock "refresh_cache-45c9e2be-eb19-4b83-b280-c9eeaddfccba" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.093742] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Preparing fetch location {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 958.094652] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Fetch image to [datastore2] OSTACK_IMG_5760532c-a109-472a-abd6-5c46be672756/OSTACK_IMG_5760532c-a109-472a-abd6-5c46be672756.vmdk {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 958.094828] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Downloading stream optimized image d9df5b42-2840-4afc-82fa-268aed408575 to [datastore2] OSTACK_IMG_5760532c-a109-472a-abd6-5c46be672756/OSTACK_IMG_5760532c-a109-472a-abd6-5c46be672756.vmdk on the data store datastore2 as vApp {{(pid=63021) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 958.095072] env[63021]: DEBUG nova.virt.vmwareapi.images [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Downloading image file data d9df5b42-2840-4afc-82fa-268aed408575 to the ESX as VM named 'OSTACK_IMG_5760532c-a109-472a-abd6-5c46be672756' {{(pid=63021) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 958.120552] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22c3a37c-ad94-4824-a1f9-3f4a3a703e02 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.143686] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-27b0e2ae-955e-4b60-bb43-06e4862f8fa3 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Doing hard reboot of VM {{(pid=63021) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 958.143976] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-79450651-483b-483c-8869-298835ce0ed2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.150741] env[63021]: DEBUG oslo_vmware.api [None req-27b0e2ae-955e-4b60-bb43-06e4862f8fa3 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 958.150741] env[63021]: value = "task-1294070" [ 958.150741] env[63021]: _type = "Task" [ 958.150741] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.160912] env[63021]: DEBUG oslo_vmware.api [None req-27b0e2ae-955e-4b60-bb43-06e4862f8fa3 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294070, 'name': ResetVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.182226] env[63021]: DEBUG oslo_vmware.rw_handles [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 958.182226] env[63021]: value = "resgroup-9" [ 958.182226] env[63021]: _type = "ResourcePool" [ 958.182226] env[63021]: }. {{(pid=63021) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 958.182596] env[63021]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-8b52a181-8411-405f-9caa-95ad71e5721c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.205340] env[63021]: DEBUG oslo_vmware.rw_handles [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lease: (returnval){ [ 958.205340] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]521f49d3-b6ec-326d-eefb-6f2388a41d9c" [ 958.205340] env[63021]: _type = "HttpNfcLease" [ 958.205340] env[63021]: } obtained for vApp import into resource pool (val){ [ 958.205340] env[63021]: value = "resgroup-9" [ 958.205340] env[63021]: _type = "ResourcePool" [ 958.205340] env[63021]: }. {{(pid=63021) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 958.206025] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Waiting for the lease: (returnval){ [ 958.206025] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]521f49d3-b6ec-326d-eefb-6f2388a41d9c" [ 958.206025] env[63021]: _type = "HttpNfcLease" [ 958.206025] env[63021]: } to be ready. {{(pid=63021) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 958.217723] env[63021]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 958.217723] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]521f49d3-b6ec-326d-eefb-6f2388a41d9c" [ 958.217723] env[63021]: _type = "HttpNfcLease" [ 958.217723] env[63021]: } is initializing. {{(pid=63021) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 958.274689] env[63021]: DEBUG oslo_concurrency.lockutils [None req-db9db38f-ae08-4a7b-8cb1-5d2701733145 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.847s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.277634] env[63021]: DEBUG oslo_concurrency.lockutils [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 8.369s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.340048] env[63021]: DEBUG oslo_vmware.api [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294068, 'name': ReconfigVM_Task, 'duration_secs': 0.926411} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.340422] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Reconfigured VM instance instance-00000058 to attach disk [datastore2] 6155c098-e4a6-47e6-b343-4a77ca90eb2e/6155c098-e4a6-47e6-b343-4a77ca90eb2e.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 958.341131] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-46715865-b777-4b33-afcb-51147d81c6f4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.347807] env[63021]: DEBUG oslo_vmware.api [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 958.347807] env[63021]: value = "task-1294072" [ 958.347807] env[63021]: _type = "Task" [ 958.347807] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.357792] env[63021]: DEBUG oslo_vmware.api [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294072, 'name': Rename_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.365362] env[63021]: DEBUG oslo_concurrency.lockutils [None req-12af6f49-90b6-46d7-b591-53c8cb0191dd tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquiring lock "refresh_cache-df856146-61ba-4de7-a5fd-d75263927163" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.365592] env[63021]: DEBUG oslo_concurrency.lockutils [None req-12af6f49-90b6-46d7-b591-53c8cb0191dd tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquired lock "refresh_cache-df856146-61ba-4de7-a5fd-d75263927163" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.365798] env[63021]: DEBUG nova.network.neutron [None req-12af6f49-90b6-46d7-b591-53c8cb0191dd tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 958.366075] env[63021]: DEBUG nova.objects.instance [None req-12af6f49-90b6-46d7-b591-53c8cb0191dd tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lazy-loading 'info_cache' on Instance uuid df856146-61ba-4de7-a5fd-d75263927163 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 958.373861] env[63021]: DEBUG oslo_vmware.api [None req-b5ea11b7-7b51-4404-9b05-ba2dae0b310c tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294069, 'name': CloneVM_Task} progress is 94%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.448543] env[63021]: DEBUG nova.compute.manager [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 958.474006] env[63021]: DEBUG nova.virt.hardware [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 958.474432] env[63021]: DEBUG nova.virt.hardware [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 958.474635] env[63021]: DEBUG nova.virt.hardware [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 958.474837] env[63021]: DEBUG nova.virt.hardware [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 958.475124] env[63021]: DEBUG nova.virt.hardware [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 958.475389] env[63021]: DEBUG nova.virt.hardware [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 958.475740] env[63021]: DEBUG nova.virt.hardware [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 958.476022] env[63021]: DEBUG nova.virt.hardware [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 958.476294] env[63021]: DEBUG nova.virt.hardware [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 958.476531] env[63021]: DEBUG nova.virt.hardware [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 958.476827] env[63021]: DEBUG nova.virt.hardware [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 958.478250] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c969f2a7-5a8f-41f8-ba38-0f39002219e6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.487046] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cb44062-318d-4f72-ad53-dd469a3c3c72 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.544340] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "67f736d9-82ec-448e-8b72-84fdd704aa8d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.544773] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "67f736d9-82ec-448e-8b72-84fdd704aa8d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.665596] env[63021]: DEBUG oslo_vmware.api [None req-27b0e2ae-955e-4b60-bb43-06e4862f8fa3 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294070, 'name': ResetVM_Task, 'duration_secs': 0.110097} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.666197] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-27b0e2ae-955e-4b60-bb43-06e4862f8fa3 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Did hard reboot of VM {{(pid=63021) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 958.666496] env[63021]: DEBUG nova.compute.manager [None req-27b0e2ae-955e-4b60-bb43-06e4862f8fa3 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 958.667381] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-536d1de3-d085-4002-a19e-f11665ea50a3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.714544] env[63021]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 958.714544] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]521f49d3-b6ec-326d-eefb-6f2388a41d9c" [ 958.714544] env[63021]: _type = "HttpNfcLease" [ 958.714544] env[63021]: } is initializing. {{(pid=63021) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 958.747317] env[63021]: DEBUG nova.compute.manager [req-75bed31e-e4ae-4475-8494-e6d8226c29b1 req-bd25f2dc-44a5-4f6f-b2ea-bd96a069040e service nova] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Received event network-vif-plugged-a8e61c39-aea7-48e6-9798-cbf508c50425 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 958.747652] env[63021]: DEBUG oslo_concurrency.lockutils [req-75bed31e-e4ae-4475-8494-e6d8226c29b1 req-bd25f2dc-44a5-4f6f-b2ea-bd96a069040e service nova] Acquiring lock "7df37eb7-ffc1-4b8e-9aba-caaf3c5338df-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.748010] env[63021]: DEBUG oslo_concurrency.lockutils [req-75bed31e-e4ae-4475-8494-e6d8226c29b1 req-bd25f2dc-44a5-4f6f-b2ea-bd96a069040e service nova] Lock "7df37eb7-ffc1-4b8e-9aba-caaf3c5338df-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.748245] env[63021]: DEBUG oslo_concurrency.lockutils [req-75bed31e-e4ae-4475-8494-e6d8226c29b1 req-bd25f2dc-44a5-4f6f-b2ea-bd96a069040e service nova] Lock "7df37eb7-ffc1-4b8e-9aba-caaf3c5338df-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.748520] env[63021]: DEBUG nova.compute.manager [req-75bed31e-e4ae-4475-8494-e6d8226c29b1 req-bd25f2dc-44a5-4f6f-b2ea-bd96a069040e service nova] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] No waiting events found dispatching network-vif-plugged-a8e61c39-aea7-48e6-9798-cbf508c50425 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 958.748802] env[63021]: WARNING nova.compute.manager [req-75bed31e-e4ae-4475-8494-e6d8226c29b1 req-bd25f2dc-44a5-4f6f-b2ea-bd96a069040e service nova] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Received unexpected event network-vif-plugged-a8e61c39-aea7-48e6-9798-cbf508c50425 for instance with vm_state building and task_state spawning. [ 958.782558] env[63021]: INFO nova.compute.claims [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 958.793720] env[63021]: DEBUG oslo_concurrency.lockutils [None req-db9db38f-ae08-4a7b-8cb1-5d2701733145 tempest-ServerActionsV293TestJSON-879144492 tempest-ServerActionsV293TestJSON-879144492-project-member] Lock "80b857c4-d9dd-4483-970f-a0f5f4bcf173" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.154s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.854395] env[63021]: DEBUG nova.network.neutron [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Successfully updated port: a8e61c39-aea7-48e6-9798-cbf508c50425 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 958.861220] env[63021]: DEBUG oslo_vmware.api [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294072, 'name': Rename_Task, 'duration_secs': 0.166085} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.861220] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 958.861220] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3c6de0e2-a8f4-4718-b83d-54f26b7860ce {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.869539] env[63021]: DEBUG nova.objects.base [None req-12af6f49-90b6-46d7-b591-53c8cb0191dd tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=63021) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 958.875350] env[63021]: DEBUG oslo_vmware.api [None req-b5ea11b7-7b51-4404-9b05-ba2dae0b310c tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294069, 'name': CloneVM_Task} progress is 95%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.876462] env[63021]: DEBUG oslo_vmware.api [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 958.876462] env[63021]: value = "task-1294073" [ 958.876462] env[63021]: _type = "Task" [ 958.876462] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.885257] env[63021]: DEBUG oslo_vmware.api [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294073, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.047129] env[63021]: DEBUG nova.compute.manager [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 959.181858] env[63021]: DEBUG oslo_concurrency.lockutils [None req-27b0e2ae-955e-4b60-bb43-06e4862f8fa3 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "c328d2f7-3398-4f25-b11c-f464be7af8a1" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.231s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.215836] env[63021]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 959.215836] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]521f49d3-b6ec-326d-eefb-6f2388a41d9c" [ 959.215836] env[63021]: _type = "HttpNfcLease" [ 959.215836] env[63021]: } is initializing. {{(pid=63021) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 959.289124] env[63021]: INFO nova.compute.resource_tracker [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Updating resource usage from migration 3a372abc-b37d-4af2-9053-e4a6f57a8bbc [ 959.356491] env[63021]: DEBUG oslo_concurrency.lockutils [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Acquiring lock "refresh_cache-7df37eb7-ffc1-4b8e-9aba-caaf3c5338df" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.356682] env[63021]: DEBUG oslo_concurrency.lockutils [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Acquired lock "refresh_cache-7df37eb7-ffc1-4b8e-9aba-caaf3c5338df" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.356785] env[63021]: DEBUG nova.network.neutron [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 959.374505] env[63021]: DEBUG oslo_vmware.api [None req-b5ea11b7-7b51-4404-9b05-ba2dae0b310c tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294069, 'name': CloneVM_Task, 'duration_secs': 1.171532} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.374704] env[63021]: INFO nova.virt.vmwareapi.vmops [None req-b5ea11b7-7b51-4404-9b05-ba2dae0b310c tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Created linked-clone VM from snapshot [ 959.375491] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-966dd5ce-79e7-441a-968a-83706d37618c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.389148] env[63021]: DEBUG nova.virt.vmwareapi.images [None req-b5ea11b7-7b51-4404-9b05-ba2dae0b310c tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Uploading image 66b81519-987f-4d99-8b0a-a8e5f28c3bc9 {{(pid=63021) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 959.393854] env[63021]: DEBUG oslo_vmware.api [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294073, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.401545] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5ea11b7-7b51-4404-9b05-ba2dae0b310c tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Destroying the VM {{(pid=63021) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 959.401818] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-8f688291-000a-4e22-a595-5e31fe18fcc2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.408461] env[63021]: DEBUG oslo_vmware.api [None req-b5ea11b7-7b51-4404-9b05-ba2dae0b310c tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 959.408461] env[63021]: value = "task-1294074" [ 959.408461] env[63021]: _type = "Task" [ 959.408461] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.421664] env[63021]: DEBUG oslo_vmware.api [None req-b5ea11b7-7b51-4404-9b05-ba2dae0b310c tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294074, 'name': Destroy_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.498229] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7b91024-1422-4755-9e08-ff0e02494b7e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.507035] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04173345-6894-40aa-9b1b-e193660ea45c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.540974] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-886d372b-ef0f-461b-a543-4727361775a8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.550915] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7fef5b2-3dab-4949-b9b0-ac3ddc153832 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.567469] env[63021]: DEBUG nova.compute.provider_tree [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 959.577706] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.619987] env[63021]: DEBUG nova.network.neutron [None req-12af6f49-90b6-46d7-b591-53c8cb0191dd tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Updating instance_info_cache with network_info: [{"id": "a43dcfd2-397e-44fe-bd1c-b6116c25a28a", "address": "fa:16:3e:c8:73:42", "network": {"id": "7258df3e-b1eb-4da6-ab52-5dadd2f87b51", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1084458843-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.144", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5661c44b534b4a7e9452e914fc80d66e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4f04accd-4a9f-41a1-b327-c96095e1ade4", "external-id": "nsx-vlan-transportzone-947", "segmentation_id": 947, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa43dcfd2-39", "ovs_interfaceid": "a43dcfd2-397e-44fe-bd1c-b6116c25a28a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.715877] env[63021]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 959.715877] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]521f49d3-b6ec-326d-eefb-6f2388a41d9c" [ 959.715877] env[63021]: _type = "HttpNfcLease" [ 959.715877] env[63021]: } is initializing. {{(pid=63021) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 959.889441] env[63021]: DEBUG oslo_vmware.api [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294073, 'name': PowerOnVM_Task, 'duration_secs': 0.625004} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.890292] env[63021]: DEBUG nova.network.neutron [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 959.892099] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 959.892320] env[63021]: INFO nova.compute.manager [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Took 8.48 seconds to spawn the instance on the hypervisor. [ 959.892503] env[63021]: DEBUG nova.compute.manager [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 959.893327] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd939f1e-9fbc-498c-ac16-1af1f8ec585f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.920987] env[63021]: DEBUG oslo_vmware.api [None req-b5ea11b7-7b51-4404-9b05-ba2dae0b310c tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294074, 'name': Destroy_Task} progress is 33%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.031500] env[63021]: DEBUG nova.network.neutron [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Updating instance_info_cache with network_info: [{"id": "a8e61c39-aea7-48e6-9798-cbf508c50425", "address": "fa:16:3e:0a:e0:d7", "network": {"id": "19ef4980-f485-4334-a6c7-0a73516f53a9", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-636108369-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "efc7381b5cb942bf9ef93d6504d783e6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8e61c39-ae", "ovs_interfaceid": "a8e61c39-aea7-48e6-9798-cbf508c50425", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.070990] env[63021]: DEBUG nova.scheduler.client.report [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 960.122710] env[63021]: DEBUG oslo_concurrency.lockutils [None req-12af6f49-90b6-46d7-b591-53c8cb0191dd tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Releasing lock "refresh_cache-df856146-61ba-4de7-a5fd-d75263927163" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.216060] env[63021]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 960.216060] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]521f49d3-b6ec-326d-eefb-6f2388a41d9c" [ 960.216060] env[63021]: _type = "HttpNfcLease" [ 960.216060] env[63021]: } is ready. {{(pid=63021) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 960.216423] env[63021]: DEBUG oslo_vmware.rw_handles [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 960.216423] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]521f49d3-b6ec-326d-eefb-6f2388a41d9c" [ 960.216423] env[63021]: _type = "HttpNfcLease" [ 960.216423] env[63021]: }. {{(pid=63021) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 960.217258] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d8b267e-7e4b-4a0f-a9ea-39e6e4c84536 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.224726] env[63021]: DEBUG oslo_vmware.rw_handles [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d0e470-5b20-3a37-bbf4-7adea4000ec9/disk-0.vmdk from lease info. {{(pid=63021) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 960.224928] env[63021]: DEBUG oslo_vmware.rw_handles [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Creating HTTP connection to write to file with size = 31667200 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d0e470-5b20-3a37-bbf4-7adea4000ec9/disk-0.vmdk. {{(pid=63021) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 960.291168] env[63021]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-765d0c6d-8aab-4655-9ccf-c449e2d1c0b4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.412185] env[63021]: INFO nova.compute.manager [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Took 18.63 seconds to build instance. [ 960.422468] env[63021]: DEBUG oslo_vmware.api [None req-b5ea11b7-7b51-4404-9b05-ba2dae0b310c tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294074, 'name': Destroy_Task, 'duration_secs': 0.672907} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.422798] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-b5ea11b7-7b51-4404-9b05-ba2dae0b310c tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Destroyed the VM [ 960.423088] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b5ea11b7-7b51-4404-9b05-ba2dae0b310c tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Deleting Snapshot of the VM instance {{(pid=63021) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 960.423378] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-79914e77-9452-440b-9cd0-99bce0a9e64f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.431186] env[63021]: DEBUG oslo_vmware.api [None req-b5ea11b7-7b51-4404-9b05-ba2dae0b310c tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 960.431186] env[63021]: value = "task-1294075" [ 960.431186] env[63021]: _type = "Task" [ 960.431186] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.442513] env[63021]: DEBUG oslo_vmware.api [None req-b5ea11b7-7b51-4404-9b05-ba2dae0b310c tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294075, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.534345] env[63021]: DEBUG oslo_concurrency.lockutils [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Releasing lock "refresh_cache-7df37eb7-ffc1-4b8e-9aba-caaf3c5338df" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.534699] env[63021]: DEBUG nova.compute.manager [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Instance network_info: |[{"id": "a8e61c39-aea7-48e6-9798-cbf508c50425", "address": "fa:16:3e:0a:e0:d7", "network": {"id": "19ef4980-f485-4334-a6c7-0a73516f53a9", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-636108369-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "efc7381b5cb942bf9ef93d6504d783e6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8e61c39-ae", "ovs_interfaceid": "a8e61c39-aea7-48e6-9798-cbf508c50425", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 960.535228] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0a:e0:d7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '74e6f6e0-95e6-4531-99e9-0e78350fb655', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a8e61c39-aea7-48e6-9798-cbf508c50425', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 960.543239] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Creating folder: Project (efc7381b5cb942bf9ef93d6504d783e6). Parent ref: group-v277447. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 960.546557] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-01f0f57b-3d89-4e22-946f-9adb67e56f9e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.559289] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Created folder: Project (efc7381b5cb942bf9ef93d6504d783e6) in parent group-v277447. [ 960.559509] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Creating folder: Instances. Parent ref: group-v277567. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 960.559780] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-27e71960-34e6-4f5b-b553-f52c774ce3a6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.571383] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Created folder: Instances in parent group-v277567. [ 960.571659] env[63021]: DEBUG oslo.service.loopingcall [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 960.571881] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 960.572165] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6333c139-15b5-4102-8a6c-6c3f8729ee35 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.590940] env[63021]: DEBUG oslo_concurrency.lockutils [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.313s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.591196] env[63021]: INFO nova.compute.manager [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Migrating [ 960.601164] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.802s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.602803] env[63021]: INFO nova.compute.claims [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 960.618753] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 960.618753] env[63021]: value = "task-1294078" [ 960.618753] env[63021]: _type = "Task" [ 960.618753] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.628294] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294078, 'name': CreateVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.631345] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-12af6f49-90b6-46d7-b591-53c8cb0191dd tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 960.631672] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-16f154cf-da12-482c-a5cb-94d38f9041f7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.640185] env[63021]: DEBUG oslo_vmware.api [None req-12af6f49-90b6-46d7-b591-53c8cb0191dd tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 960.640185] env[63021]: value = "task-1294079" [ 960.640185] env[63021]: _type = "Task" [ 960.640185] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.651102] env[63021]: DEBUG oslo_vmware.api [None req-12af6f49-90b6-46d7-b591-53c8cb0191dd tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294079, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.775443] env[63021]: DEBUG nova.compute.manager [req-4bc86bc3-f369-46ad-917b-978c9676d064 req-654c6563-ed79-42de-9bb4-39e8f1c8bcb6 service nova] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Received event network-changed-a8e61c39-aea7-48e6-9798-cbf508c50425 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 960.775650] env[63021]: DEBUG nova.compute.manager [req-4bc86bc3-f369-46ad-917b-978c9676d064 req-654c6563-ed79-42de-9bb4-39e8f1c8bcb6 service nova] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Refreshing instance network info cache due to event network-changed-a8e61c39-aea7-48e6-9798-cbf508c50425. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 960.776059] env[63021]: DEBUG oslo_concurrency.lockutils [req-4bc86bc3-f369-46ad-917b-978c9676d064 req-654c6563-ed79-42de-9bb4-39e8f1c8bcb6 service nova] Acquiring lock "refresh_cache-7df37eb7-ffc1-4b8e-9aba-caaf3c5338df" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.776267] env[63021]: DEBUG oslo_concurrency.lockutils [req-4bc86bc3-f369-46ad-917b-978c9676d064 req-654c6563-ed79-42de-9bb4-39e8f1c8bcb6 service nova] Acquired lock "refresh_cache-7df37eb7-ffc1-4b8e-9aba-caaf3c5338df" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.776470] env[63021]: DEBUG nova.network.neutron [req-4bc86bc3-f369-46ad-917b-978c9676d064 req-654c6563-ed79-42de-9bb4-39e8f1c8bcb6 service nova] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Refreshing network info cache for port a8e61c39-aea7-48e6-9798-cbf508c50425 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 960.918332] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85d7292f-44f1-41b9-b04b-200332f5a783 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "6155c098-e4a6-47e6-b343-4a77ca90eb2e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.150s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.944699] env[63021]: DEBUG oslo_vmware.api [None req-b5ea11b7-7b51-4404-9b05-ba2dae0b310c tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294075, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.121136] env[63021]: DEBUG oslo_concurrency.lockutils [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "refresh_cache-207c8197-6191-4837-8ead-8c24aa5b35bc" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.121136] env[63021]: DEBUG oslo_concurrency.lockutils [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquired lock "refresh_cache-207c8197-6191-4837-8ead-8c24aa5b35bc" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.121136] env[63021]: DEBUG nova.network.neutron [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 961.136035] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294078, 'name': CreateVM_Task, 'duration_secs': 0.396826} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.138268] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 961.139180] env[63021]: DEBUG oslo_concurrency.lockutils [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.139353] env[63021]: DEBUG oslo_concurrency.lockutils [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.139683] env[63021]: DEBUG oslo_concurrency.lockutils [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 961.139945] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5bcb8250-65fb-4fa1-bc25-b9c5327ca189 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.148971] env[63021]: DEBUG oslo_vmware.api [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Waiting for the task: (returnval){ [ 961.148971] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5291d4d6-220e-6293-9bc2-cc85042eca1e" [ 961.148971] env[63021]: _type = "Task" [ 961.148971] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.157921] env[63021]: DEBUG oslo_vmware.api [None req-12af6f49-90b6-46d7-b591-53c8cb0191dd tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294079, 'name': PowerOnVM_Task, 'duration_secs': 0.470187} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.158619] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-12af6f49-90b6-46d7-b591-53c8cb0191dd tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 961.158787] env[63021]: DEBUG nova.compute.manager [None req-12af6f49-90b6-46d7-b591-53c8cb0191dd tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 961.159715] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45c5d2f3-9d9a-4662-b71d-f4d353ab8daf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.166677] env[63021]: DEBUG oslo_vmware.api [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5291d4d6-220e-6293-9bc2-cc85042eca1e, 'name': SearchDatastore_Task, 'duration_secs': 0.010531} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.169864] env[63021]: DEBUG oslo_concurrency.lockutils [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.170221] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 961.170490] env[63021]: DEBUG oslo_concurrency.lockutils [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.170688] env[63021]: DEBUG oslo_concurrency.lockutils [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.170896] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 961.172939] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-24f16ca4-b0d8-4c62-81f8-9039b30a60d4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.184703] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 961.184703] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 961.185541] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be090b03-ffc6-4684-88f0-d559d1ff5ed7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.191009] env[63021]: DEBUG oslo_vmware.api [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Waiting for the task: (returnval){ [ 961.191009] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52a65cd2-fda1-64b1-d81e-6b6274c9f1b8" [ 961.191009] env[63021]: _type = "Task" [ 961.191009] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.199406] env[63021]: DEBUG oslo_vmware.api [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52a65cd2-fda1-64b1-d81e-6b6274c9f1b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.443464] env[63021]: DEBUG oslo_vmware.api [None req-b5ea11b7-7b51-4404-9b05-ba2dae0b310c tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294075, 'name': RemoveSnapshot_Task, 'duration_secs': 0.588928} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.443901] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b5ea11b7-7b51-4404-9b05-ba2dae0b310c tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Deleted Snapshot of the VM instance {{(pid=63021) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 961.450030] env[63021]: DEBUG oslo_vmware.rw_handles [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Completed reading data from the image iterator. {{(pid=63021) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 961.450030] env[63021]: DEBUG oslo_vmware.rw_handles [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d0e470-5b20-3a37-bbf4-7adea4000ec9/disk-0.vmdk. {{(pid=63021) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 961.450030] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c84bf609-e8aa-44b6-8d27-8b98899048a9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.458680] env[63021]: DEBUG oslo_vmware.rw_handles [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d0e470-5b20-3a37-bbf4-7adea4000ec9/disk-0.vmdk is in state: ready. {{(pid=63021) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 961.458943] env[63021]: DEBUG oslo_vmware.rw_handles [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d0e470-5b20-3a37-bbf4-7adea4000ec9/disk-0.vmdk. {{(pid=63021) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 961.460625] env[63021]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-5c7cf47b-c004-40fc-96bb-9bb674fd1203 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.538207] env[63021]: DEBUG nova.network.neutron [req-4bc86bc3-f369-46ad-917b-978c9676d064 req-654c6563-ed79-42de-9bb4-39e8f1c8bcb6 service nova] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Updated VIF entry in instance network info cache for port a8e61c39-aea7-48e6-9798-cbf508c50425. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 961.538674] env[63021]: DEBUG nova.network.neutron [req-4bc86bc3-f369-46ad-917b-978c9676d064 req-654c6563-ed79-42de-9bb4-39e8f1c8bcb6 service nova] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Updating instance_info_cache with network_info: [{"id": "a8e61c39-aea7-48e6-9798-cbf508c50425", "address": "fa:16:3e:0a:e0:d7", "network": {"id": "19ef4980-f485-4334-a6c7-0a73516f53a9", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-636108369-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "efc7381b5cb942bf9ef93d6504d783e6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8e61c39-ae", "ovs_interfaceid": "a8e61c39-aea7-48e6-9798-cbf508c50425", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.707339] env[63021]: DEBUG oslo_vmware.api [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52a65cd2-fda1-64b1-d81e-6b6274c9f1b8, 'name': SearchDatastore_Task, 'duration_secs': 0.011458} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.708055] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11f48739-0d35-4943-b238-091f0c7ad86f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.715952] env[63021]: DEBUG oslo_vmware.api [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Waiting for the task: (returnval){ [ 961.715952] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52a381a8-f7ce-6ccb-a52f-b595cb6ba411" [ 961.715952] env[63021]: _type = "Task" [ 961.715952] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.728072] env[63021]: DEBUG oslo_vmware.api [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52a381a8-f7ce-6ccb-a52f-b595cb6ba411, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.815665] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2afa05bc-fa38-40f4-956d-2aea44ef021f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.823291] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15fb9dd7-82be-44b5-985d-99a4d5b690e5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.855850] env[63021]: DEBUG nova.network.neutron [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Updating instance_info_cache with network_info: [{"id": "a7ecacae-44cb-4b30-892f-d0f5ef3ad9d5", "address": "fa:16:3e:b7:1c:44", "network": {"id": "708d0dc3-c651-4001-9638-1d3538ff76c9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-983471056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "805af233efa147b3982fb794cd9e6f73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7ecacae-44", "ovs_interfaceid": "a7ecacae-44cb-4b30-892f-d0f5ef3ad9d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.857608] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd6a13a6-8f3c-408e-9d6e-84e4fd20baa2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.869116] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8146d2f-923a-448f-b537-f68d03bbbc2b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.883638] env[63021]: DEBUG nova.compute.provider_tree [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 961.943046] env[63021]: DEBUG oslo_vmware.rw_handles [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d0e470-5b20-3a37-bbf4-7adea4000ec9/disk-0.vmdk. {{(pid=63021) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 961.943387] env[63021]: INFO nova.virt.vmwareapi.images [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Downloaded image file data d9df5b42-2840-4afc-82fa-268aed408575 [ 961.944211] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2671d93c-4905-4a8d-94fd-498a1e93ef20 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.949929] env[63021]: DEBUG nova.compute.manager [req-1ca8dd2c-31c7-43f4-8949-2b765f6e8760 req-ea33cc84-068a-4a1d-9302-b5dfcb8dc40f service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Received event network-changed-eab7f4fe-4832-40e4-aed2-0e931d91a4d7 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 961.950160] env[63021]: DEBUG nova.compute.manager [req-1ca8dd2c-31c7-43f4-8949-2b765f6e8760 req-ea33cc84-068a-4a1d-9302-b5dfcb8dc40f service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Refreshing instance network info cache due to event network-changed-eab7f4fe-4832-40e4-aed2-0e931d91a4d7. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 961.950437] env[63021]: DEBUG oslo_concurrency.lockutils [req-1ca8dd2c-31c7-43f4-8949-2b765f6e8760 req-ea33cc84-068a-4a1d-9302-b5dfcb8dc40f service nova] Acquiring lock "refresh_cache-2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.950525] env[63021]: DEBUG oslo_concurrency.lockutils [req-1ca8dd2c-31c7-43f4-8949-2b765f6e8760 req-ea33cc84-068a-4a1d-9302-b5dfcb8dc40f service nova] Acquired lock "refresh_cache-2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.950688] env[63021]: DEBUG nova.network.neutron [req-1ca8dd2c-31c7-43f4-8949-2b765f6e8760 req-ea33cc84-068a-4a1d-9302-b5dfcb8dc40f service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Refreshing network info cache for port eab7f4fe-4832-40e4-aed2-0e931d91a4d7 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 961.963758] env[63021]: WARNING nova.compute.manager [None req-b5ea11b7-7b51-4404-9b05-ba2dae0b310c tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Image not found during snapshot: nova.exception.ImageNotFound: Image 66b81519-987f-4d99-8b0a-a8e5f28c3bc9 could not be found. [ 961.967738] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4719834e-2ca3-48a5-ada2-c6b9e3847db9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.042925] env[63021]: DEBUG oslo_concurrency.lockutils [req-4bc86bc3-f369-46ad-917b-978c9676d064 req-654c6563-ed79-42de-9bb4-39e8f1c8bcb6 service nova] Releasing lock "refresh_cache-7df37eb7-ffc1-4b8e-9aba-caaf3c5338df" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.063996] env[63021]: INFO nova.virt.vmwareapi.images [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] The imported VM was unregistered [ 962.067458] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Caching image {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 962.067736] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Creating directory with path [datastore2] devstack-image-cache_base/d9df5b42-2840-4afc-82fa-268aed408575 {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 962.068063] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e3c11cef-3b2e-4590-9d10-f94d1c1a570b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.098862] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Created directory with path [datastore2] devstack-image-cache_base/d9df5b42-2840-4afc-82fa-268aed408575 {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 962.099102] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_5760532c-a109-472a-abd6-5c46be672756/OSTACK_IMG_5760532c-a109-472a-abd6-5c46be672756.vmdk to [datastore2] devstack-image-cache_base/d9df5b42-2840-4afc-82fa-268aed408575/d9df5b42-2840-4afc-82fa-268aed408575.vmdk. {{(pid=63021) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 962.099389] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-02579ada-e1d3-4dca-8bf6-a240cae4d7c8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.108497] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Waiting for the task: (returnval){ [ 962.108497] env[63021]: value = "task-1294081" [ 962.108497] env[63021]: _type = "Task" [ 962.108497] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.116841] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1294081, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.208864] env[63021]: DEBUG nova.network.neutron [req-1ca8dd2c-31c7-43f4-8949-2b765f6e8760 req-ea33cc84-068a-4a1d-9302-b5dfcb8dc40f service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Updated VIF entry in instance network info cache for port eab7f4fe-4832-40e4-aed2-0e931d91a4d7. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 962.209274] env[63021]: DEBUG nova.network.neutron [req-1ca8dd2c-31c7-43f4-8949-2b765f6e8760 req-ea33cc84-068a-4a1d-9302-b5dfcb8dc40f service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Updating instance_info_cache with network_info: [{"id": "eab7f4fe-4832-40e4-aed2-0e931d91a4d7", "address": "fa:16:3e:9c:24:88", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeab7f4fe-48", "ovs_interfaceid": "eab7f4fe-4832-40e4-aed2-0e931d91a4d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.229045] env[63021]: DEBUG oslo_vmware.api [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52a381a8-f7ce-6ccb-a52f-b595cb6ba411, 'name': SearchDatastore_Task, 'duration_secs': 0.030492} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.229322] env[63021]: DEBUG oslo_concurrency.lockutils [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.229606] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df/7df37eb7-ffc1-4b8e-9aba-caaf3c5338df.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 962.229878] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6c74690d-d11a-49a6-ab59-3c411921794a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.235844] env[63021]: DEBUG oslo_vmware.api [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Waiting for the task: (returnval){ [ 962.235844] env[63021]: value = "task-1294082" [ 962.235844] env[63021]: _type = "Task" [ 962.235844] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.244158] env[63021]: DEBUG oslo_vmware.api [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Task: {'id': task-1294082, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.365692] env[63021]: DEBUG oslo_concurrency.lockutils [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Releasing lock "refresh_cache-207c8197-6191-4837-8ead-8c24aa5b35bc" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.386348] env[63021]: DEBUG nova.scheduler.client.report [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 962.620342] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1294081, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.712920] env[63021]: DEBUG oslo_concurrency.lockutils [req-1ca8dd2c-31c7-43f4-8949-2b765f6e8760 req-ea33cc84-068a-4a1d-9302-b5dfcb8dc40f service nova] Releasing lock "refresh_cache-2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.752420] env[63021]: DEBUG oslo_vmware.api [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Task: {'id': task-1294082, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.893361] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.292s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.893874] env[63021]: DEBUG nova.compute.manager [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 962.900250] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.319s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.900510] env[63021]: DEBUG nova.objects.instance [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Lazy-loading 'resources' on Instance uuid a52e4cb5-d0ea-4698-9955-753626762a02 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 963.123959] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1294081, 'name': MoveVirtualDisk_Task} progress is 26%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.170241] env[63021]: DEBUG nova.compute.manager [req-4c9ffcb5-29d1-4269-8ca5-9139eaf78757 req-92b6d266-ca2c-4900-bd50-3fdba53699da service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Received event network-changed-3752044a-9145-469d-8edb-6255d7d0f928 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 963.170389] env[63021]: DEBUG nova.compute.manager [req-4c9ffcb5-29d1-4269-8ca5-9139eaf78757 req-92b6d266-ca2c-4900-bd50-3fdba53699da service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Refreshing instance network info cache due to event network-changed-3752044a-9145-469d-8edb-6255d7d0f928. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 963.170633] env[63021]: DEBUG oslo_concurrency.lockutils [req-4c9ffcb5-29d1-4269-8ca5-9139eaf78757 req-92b6d266-ca2c-4900-bd50-3fdba53699da service nova] Acquiring lock "refresh_cache-6155c098-e4a6-47e6-b343-4a77ca90eb2e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.170793] env[63021]: DEBUG oslo_concurrency.lockutils [req-4c9ffcb5-29d1-4269-8ca5-9139eaf78757 req-92b6d266-ca2c-4900-bd50-3fdba53699da service nova] Acquired lock "refresh_cache-6155c098-e4a6-47e6-b343-4a77ca90eb2e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.171025] env[63021]: DEBUG nova.network.neutron [req-4c9ffcb5-29d1-4269-8ca5-9139eaf78757 req-92b6d266-ca2c-4900-bd50-3fdba53699da service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Refreshing network info cache for port 3752044a-9145-469d-8edb-6255d7d0f928 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 963.224343] env[63021]: DEBUG oslo_concurrency.lockutils [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "919887c8-171d-48fe-8c6b-2abf5f55d6ab" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.224851] env[63021]: DEBUG oslo_concurrency.lockutils [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "919887c8-171d-48fe-8c6b-2abf5f55d6ab" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.224851] env[63021]: DEBUG oslo_concurrency.lockutils [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "919887c8-171d-48fe-8c6b-2abf5f55d6ab-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.224991] env[63021]: DEBUG oslo_concurrency.lockutils [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "919887c8-171d-48fe-8c6b-2abf5f55d6ab-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.225140] env[63021]: DEBUG oslo_concurrency.lockutils [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "919887c8-171d-48fe-8c6b-2abf5f55d6ab-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.227488] env[63021]: INFO nova.compute.manager [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Terminating instance [ 963.229500] env[63021]: DEBUG nova.compute.manager [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 963.229705] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 963.230565] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0da550cc-5075-491a-9a59-1a4961d68b82 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.243085] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 963.243761] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-865d4fa1-afec-4944-942b-86c152532849 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.250641] env[63021]: DEBUG oslo_vmware.api [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Task: {'id': task-1294082, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.257628] env[63021]: DEBUG oslo_vmware.api [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 963.257628] env[63021]: value = "task-1294083" [ 963.257628] env[63021]: _type = "Task" [ 963.257628] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.270684] env[63021]: DEBUG oslo_vmware.api [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294083, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.404097] env[63021]: DEBUG nova.compute.utils [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 963.408688] env[63021]: DEBUG nova.compute.manager [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 963.408914] env[63021]: DEBUG nova.network.neutron [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 963.455845] env[63021]: DEBUG nova.policy [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b427f7d1317341f38754a2b324f4f5aa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e238d5c1668540d2a9fffd3fd832d9b8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 963.630870] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1294081, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.638016] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-734ba79e-2bf6-4d65-8218-075f077b5dee {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.650165] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f6b5c7-af19-4b38-9f0a-ad84388befd0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.693187] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f54216a1-d01f-4508-948c-2eb121ef0137 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.706529] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-998d9a8d-b174-4b2e-9aed-d7a4de6972e7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.734207] env[63021]: DEBUG nova.compute.provider_tree [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 963.757978] env[63021]: DEBUG oslo_vmware.api [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Task: {'id': task-1294082, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.433867} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.758503] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df/7df37eb7-ffc1-4b8e-9aba-caaf3c5338df.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 963.758713] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 963.762533] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9677b0ec-a214-4caf-b4d6-f437ca3bade9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.772257] env[63021]: DEBUG oslo_vmware.api [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Waiting for the task: (returnval){ [ 963.772257] env[63021]: value = "task-1294084" [ 963.772257] env[63021]: _type = "Task" [ 963.772257] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.777921] env[63021]: DEBUG oslo_vmware.api [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294083, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.789559] env[63021]: DEBUG oslo_vmware.api [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Task: {'id': task-1294084, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.814852] env[63021]: DEBUG nova.network.neutron [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Successfully created port: 7aec4a00-7fef-4c0f-a168-2cfc26dbe0be {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 963.882497] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91fc402b-57ac-46f9-8f7c-f7a927ea1197 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.911421] env[63021]: DEBUG nova.compute.manager [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 963.914562] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Updating instance '207c8197-6191-4837-8ead-8c24aa5b35bc' progress to 0 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 964.125597] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1294081, 'name': MoveVirtualDisk_Task} progress is 60%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.132187] env[63021]: DEBUG nova.compute.manager [req-486f982a-2a8c-41fb-bc23-0c5f979d7091 req-6f484427-c479-42ea-aa75-5f05cc8e5c6c service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Received event network-changed-3752044a-9145-469d-8edb-6255d7d0f928 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 964.132434] env[63021]: DEBUG nova.compute.manager [req-486f982a-2a8c-41fb-bc23-0c5f979d7091 req-6f484427-c479-42ea-aa75-5f05cc8e5c6c service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Refreshing instance network info cache due to event network-changed-3752044a-9145-469d-8edb-6255d7d0f928. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 964.132746] env[63021]: DEBUG oslo_concurrency.lockutils [req-486f982a-2a8c-41fb-bc23-0c5f979d7091 req-6f484427-c479-42ea-aa75-5f05cc8e5c6c service nova] Acquiring lock "refresh_cache-6155c098-e4a6-47e6-b343-4a77ca90eb2e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.151271] env[63021]: DEBUG nova.network.neutron [req-4c9ffcb5-29d1-4269-8ca5-9139eaf78757 req-92b6d266-ca2c-4900-bd50-3fdba53699da service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Updated VIF entry in instance network info cache for port 3752044a-9145-469d-8edb-6255d7d0f928. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 964.151271] env[63021]: DEBUG nova.network.neutron [req-4c9ffcb5-29d1-4269-8ca5-9139eaf78757 req-92b6d266-ca2c-4900-bd50-3fdba53699da service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Updating instance_info_cache with network_info: [{"id": "3752044a-9145-469d-8edb-6255d7d0f928", "address": "fa:16:3e:fe:57:1f", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3752044a-91", "ovs_interfaceid": "3752044a-9145-469d-8edb-6255d7d0f928", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.240075] env[63021]: DEBUG nova.scheduler.client.report [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 964.273107] env[63021]: DEBUG oslo_vmware.api [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294083, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.289163] env[63021]: DEBUG oslo_vmware.api [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Task: {'id': task-1294084, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.251951} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.289468] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 964.290381] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-564fa0dd-3c74-4417-95bd-9d2685e42cf3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.320267] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df/7df37eb7-ffc1-4b8e-9aba-caaf3c5338df.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 964.321349] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1839ccb3-829f-45a9-8ed7-24e1bcb3571c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.346354] env[63021]: DEBUG oslo_vmware.api [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Waiting for the task: (returnval){ [ 964.346354] env[63021]: value = "task-1294085" [ 964.346354] env[63021]: _type = "Task" [ 964.346354] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.359369] env[63021]: DEBUG oslo_vmware.api [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Task: {'id': task-1294085, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.426059] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 964.429581] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3473a00c-5bc9-4ed4-86cc-b9679ab75bdf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.440477] env[63021]: DEBUG oslo_vmware.api [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 964.440477] env[63021]: value = "task-1294086" [ 964.440477] env[63021]: _type = "Task" [ 964.440477] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.456404] env[63021]: DEBUG oslo_vmware.api [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294086, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.627337] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1294081, 'name': MoveVirtualDisk_Task} progress is 80%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.654016] env[63021]: DEBUG oslo_concurrency.lockutils [req-4c9ffcb5-29d1-4269-8ca5-9139eaf78757 req-92b6d266-ca2c-4900-bd50-3fdba53699da service nova] Releasing lock "refresh_cache-6155c098-e4a6-47e6-b343-4a77ca90eb2e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.654635] env[63021]: DEBUG oslo_concurrency.lockutils [req-486f982a-2a8c-41fb-bc23-0c5f979d7091 req-6f484427-c479-42ea-aa75-5f05cc8e5c6c service nova] Acquired lock "refresh_cache-6155c098-e4a6-47e6-b343-4a77ca90eb2e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.654838] env[63021]: DEBUG nova.network.neutron [req-486f982a-2a8c-41fb-bc23-0c5f979d7091 req-6f484427-c479-42ea-aa75-5f05cc8e5c6c service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Refreshing network info cache for port 3752044a-9145-469d-8edb-6255d7d0f928 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 964.743719] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.843s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.746453] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.693s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.747486] env[63021]: DEBUG nova.objects.instance [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Lazy-loading 'resources' on Instance uuid ed26dff1-fed0-4baf-ad41-d14850254aec {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 964.775456] env[63021]: DEBUG oslo_vmware.api [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294083, 'name': PowerOffVM_Task, 'duration_secs': 1.161787} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.776772] env[63021]: INFO nova.scheduler.client.report [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Deleted allocations for instance a52e4cb5-d0ea-4698-9955-753626762a02 [ 964.778072] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 964.778340] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 964.781610] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ed7029fe-47c6-4eb6-8f07-667011d0c1b0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.860087] env[63021]: DEBUG oslo_vmware.api [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Task: {'id': task-1294085, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.910663] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 964.911147] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Deleting contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 964.911425] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Deleting the datastore file [datastore2] 919887c8-171d-48fe-8c6b-2abf5f55d6ab {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 964.911813] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a369b45f-6c40-4f7b-9e0f-58309ebed8e6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.924665] env[63021]: DEBUG oslo_vmware.api [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for the task: (returnval){ [ 964.924665] env[63021]: value = "task-1294088" [ 964.924665] env[63021]: _type = "Task" [ 964.924665] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.933456] env[63021]: DEBUG nova.compute.manager [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 964.943770] env[63021]: DEBUG oslo_vmware.api [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294088, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.959770] env[63021]: DEBUG oslo_vmware.api [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294086, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.971361] env[63021]: DEBUG nova.virt.hardware [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 964.971710] env[63021]: DEBUG nova.virt.hardware [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 964.971877] env[63021]: DEBUG nova.virt.hardware [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 964.972093] env[63021]: DEBUG nova.virt.hardware [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 964.972288] env[63021]: DEBUG nova.virt.hardware [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 964.972622] env[63021]: DEBUG nova.virt.hardware [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 964.972937] env[63021]: DEBUG nova.virt.hardware [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 964.973208] env[63021]: DEBUG nova.virt.hardware [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 964.973458] env[63021]: DEBUG nova.virt.hardware [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 964.973712] env[63021]: DEBUG nova.virt.hardware [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 964.973945] env[63021]: DEBUG nova.virt.hardware [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 964.975071] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3549208a-7d43-444e-a1b0-c6ebe73be133 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.989288] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2e6018d-5a12-4367-9d90-fe2eceaa96c4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.124317] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1294081, 'name': MoveVirtualDisk_Task} progress is 97%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.230320] env[63021]: DEBUG nova.compute.manager [req-0242a54b-518a-4493-81b0-49c262c7199d req-8c807ada-979c-4ea2-a477-e6fc073a0562 service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Received event network-changed-eab7f4fe-4832-40e4-aed2-0e931d91a4d7 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 965.230624] env[63021]: DEBUG nova.compute.manager [req-0242a54b-518a-4493-81b0-49c262c7199d req-8c807ada-979c-4ea2-a477-e6fc073a0562 service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Refreshing instance network info cache due to event network-changed-eab7f4fe-4832-40e4-aed2-0e931d91a4d7. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 965.230751] env[63021]: DEBUG oslo_concurrency.lockutils [req-0242a54b-518a-4493-81b0-49c262c7199d req-8c807ada-979c-4ea2-a477-e6fc073a0562 service nova] Acquiring lock "refresh_cache-2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.230857] env[63021]: DEBUG oslo_concurrency.lockutils [req-0242a54b-518a-4493-81b0-49c262c7199d req-8c807ada-979c-4ea2-a477-e6fc073a0562 service nova] Acquired lock "refresh_cache-2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.231051] env[63021]: DEBUG nova.network.neutron [req-0242a54b-518a-4493-81b0-49c262c7199d req-8c807ada-979c-4ea2-a477-e6fc073a0562 service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Refreshing network info cache for port eab7f4fe-4832-40e4-aed2-0e931d91a4d7 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 965.286839] env[63021]: DEBUG oslo_concurrency.lockutils [None req-85c47402-f345-44fd-a983-c9f3341e0ff9 tempest-ServerMetadataTestJSON-1406343555 tempest-ServerMetadataTestJSON-1406343555-project-member] Lock "a52e4cb5-d0ea-4698-9955-753626762a02" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.769s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.363999] env[63021]: DEBUG oslo_vmware.api [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Task: {'id': task-1294085, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.378229] env[63021]: DEBUG nova.network.neutron [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Successfully updated port: 7aec4a00-7fef-4c0f-a168-2cfc26dbe0be {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 965.438471] env[63021]: DEBUG oslo_vmware.api [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294088, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.455861] env[63021]: DEBUG oslo_vmware.api [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294086, 'name': PowerOffVM_Task, 'duration_secs': 0.596576} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.456521] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 965.456717] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Updating instance '207c8197-6191-4837-8ead-8c24aa5b35bc' progress to 17 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 965.492421] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f514f70-818c-45e3-beeb-f5781f6acd59 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.502737] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a122b9ad-795e-4647-836f-07f41c63509f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.534387] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afeaaf33-0097-42e5-b8d8-8634d05b0c9d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.542499] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8538ea91-e752-4ed4-91f5-0cc382025368 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.557897] env[63021]: DEBUG nova.compute.provider_tree [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 965.623904] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1294081, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.306081} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.624324] env[63021]: INFO nova.virt.vmwareapi.ds_util [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_5760532c-a109-472a-abd6-5c46be672756/OSTACK_IMG_5760532c-a109-472a-abd6-5c46be672756.vmdk to [datastore2] devstack-image-cache_base/d9df5b42-2840-4afc-82fa-268aed408575/d9df5b42-2840-4afc-82fa-268aed408575.vmdk. [ 965.624519] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Cleaning up location [datastore2] OSTACK_IMG_5760532c-a109-472a-abd6-5c46be672756 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 965.624688] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_5760532c-a109-472a-abd6-5c46be672756 {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 965.624966] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-96d570a7-6b94-43d1-8acc-bf26d8a1d4eb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.632482] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Waiting for the task: (returnval){ [ 965.632482] env[63021]: value = "task-1294089" [ 965.632482] env[63021]: _type = "Task" [ 965.632482] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.641204] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1294089, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.641678] env[63021]: DEBUG nova.network.neutron [req-486f982a-2a8c-41fb-bc23-0c5f979d7091 req-6f484427-c479-42ea-aa75-5f05cc8e5c6c service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Updated VIF entry in instance network info cache for port 3752044a-9145-469d-8edb-6255d7d0f928. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 965.642026] env[63021]: DEBUG nova.network.neutron [req-486f982a-2a8c-41fb-bc23-0c5f979d7091 req-6f484427-c479-42ea-aa75-5f05cc8e5c6c service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Updating instance_info_cache with network_info: [{"id": "3752044a-9145-469d-8edb-6255d7d0f928", "address": "fa:16:3e:fe:57:1f", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3752044a-91", "ovs_interfaceid": "3752044a-9145-469d-8edb-6255d7d0f928", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.856453] env[63021]: DEBUG oslo_vmware.api [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Task: {'id': task-1294085, 'name': ReconfigVM_Task, 'duration_secs': 1.429023} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.856767] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Reconfigured VM instance instance-00000059 to attach disk [datastore2] 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df/7df37eb7-ffc1-4b8e-9aba-caaf3c5338df.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 965.857497] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-772e8f25-4ed3-4339-86ad-b0d9c6665da1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.863344] env[63021]: DEBUG oslo_vmware.api [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Waiting for the task: (returnval){ [ 965.863344] env[63021]: value = "task-1294090" [ 965.863344] env[63021]: _type = "Task" [ 965.863344] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.873877] env[63021]: DEBUG oslo_vmware.api [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Task: {'id': task-1294090, 'name': Rename_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.878573] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "refresh_cache-cee7a86d-91e2-4b57-b6cd-29056da6a414" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.878671] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquired lock "refresh_cache-cee7a86d-91e2-4b57-b6cd-29056da6a414" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.878847] env[63021]: DEBUG nova.network.neutron [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 965.934595] env[63021]: DEBUG oslo_vmware.api [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Task: {'id': task-1294088, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.752861} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.934941] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 965.935180] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Deleted contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 965.935376] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 965.935546] env[63021]: INFO nova.compute.manager [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Took 2.71 seconds to destroy the instance on the hypervisor. [ 965.935837] env[63021]: DEBUG oslo.service.loopingcall [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 965.939025] env[63021]: DEBUG nova.compute.manager [-] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 965.939025] env[63021]: DEBUG nova.network.neutron [-] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 965.963158] env[63021]: DEBUG nova.virt.hardware [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 965.963447] env[63021]: DEBUG nova.virt.hardware [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 965.963608] env[63021]: DEBUG nova.virt.hardware [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 965.963794] env[63021]: DEBUG nova.virt.hardware [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 965.963961] env[63021]: DEBUG nova.virt.hardware [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 965.964159] env[63021]: DEBUG nova.virt.hardware [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 965.964468] env[63021]: DEBUG nova.virt.hardware [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 965.964708] env[63021]: DEBUG nova.virt.hardware [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 965.964815] env[63021]: DEBUG nova.virt.hardware [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 965.964982] env[63021]: DEBUG nova.virt.hardware [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 965.965233] env[63021]: DEBUG nova.virt.hardware [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 965.970354] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a01176de-7b4a-4f26-b598-764e63725bb6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.985794] env[63021]: DEBUG oslo_vmware.api [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 965.985794] env[63021]: value = "task-1294091" [ 965.985794] env[63021]: _type = "Task" [ 965.985794] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.994050] env[63021]: DEBUG oslo_vmware.api [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294091, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.061240] env[63021]: DEBUG nova.scheduler.client.report [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 966.143223] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1294089, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.091147} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.144114] env[63021]: DEBUG oslo_concurrency.lockutils [req-486f982a-2a8c-41fb-bc23-0c5f979d7091 req-6f484427-c479-42ea-aa75-5f05cc8e5c6c service nova] Releasing lock "refresh_cache-6155c098-e4a6-47e6-b343-4a77ca90eb2e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.144666] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 966.144666] env[63021]: DEBUG oslo_concurrency.lockutils [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d9df5b42-2840-4afc-82fa-268aed408575/d9df5b42-2840-4afc-82fa-268aed408575.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.148020] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d9df5b42-2840-4afc-82fa-268aed408575/d9df5b42-2840-4afc-82fa-268aed408575.vmdk to [datastore2] 45c9e2be-eb19-4b83-b280-c9eeaddfccba/45c9e2be-eb19-4b83-b280-c9eeaddfccba.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 966.148020] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-59a5b7b1-ff86-46f4-ab11-68492e1834fa {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.151921] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Waiting for the task: (returnval){ [ 966.151921] env[63021]: value = "task-1294092" [ 966.151921] env[63021]: _type = "Task" [ 966.151921] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.163931] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1294092, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.373462] env[63021]: DEBUG oslo_vmware.api [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Task: {'id': task-1294090, 'name': Rename_Task, 'duration_secs': 0.377147} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.373796] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 966.374032] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7cb6489e-cef3-4929-a406-94db46dce797 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.380081] env[63021]: DEBUG oslo_vmware.api [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Waiting for the task: (returnval){ [ 966.380081] env[63021]: value = "task-1294093" [ 966.380081] env[63021]: _type = "Task" [ 966.380081] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.389586] env[63021]: DEBUG oslo_vmware.api [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Task: {'id': task-1294093, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.411882] env[63021]: DEBUG nova.network.neutron [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 966.430776] env[63021]: DEBUG nova.network.neutron [req-0242a54b-518a-4493-81b0-49c262c7199d req-8c807ada-979c-4ea2-a477-e6fc073a0562 service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Updated VIF entry in instance network info cache for port eab7f4fe-4832-40e4-aed2-0e931d91a4d7. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 966.431200] env[63021]: DEBUG nova.network.neutron [req-0242a54b-518a-4493-81b0-49c262c7199d req-8c807ada-979c-4ea2-a477-e6fc073a0562 service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Updating instance_info_cache with network_info: [{"id": "eab7f4fe-4832-40e4-aed2-0e931d91a4d7", "address": "fa:16:3e:9c:24:88", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeab7f4fe-48", "ovs_interfaceid": "eab7f4fe-4832-40e4-aed2-0e931d91a4d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.496325] env[63021]: DEBUG oslo_vmware.api [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294091, 'name': ReconfigVM_Task, 'duration_secs': 0.291429} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.496702] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Updating instance '207c8197-6191-4837-8ead-8c24aa5b35bc' progress to 33 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 966.553532] env[63021]: DEBUG nova.network.neutron [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Updating instance_info_cache with network_info: [{"id": "7aec4a00-7fef-4c0f-a168-2cfc26dbe0be", "address": "fa:16:3e:f3:38:47", "network": {"id": "81917a55-4e41-4898-9f4e-4fced52b5a86", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1793837736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e238d5c1668540d2a9fffd3fd832d9b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7aec4a00-7f", "ovs_interfaceid": "7aec4a00-7fef-4c0f-a168-2cfc26dbe0be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.566409] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.820s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.569772] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.992s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.571431] env[63021]: INFO nova.compute.claims [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 966.595986] env[63021]: INFO nova.scheduler.client.report [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Deleted allocations for instance ed26dff1-fed0-4baf-ad41-d14850254aec [ 966.668114] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1294092, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.702305] env[63021]: DEBUG nova.network.neutron [-] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.894719] env[63021]: DEBUG oslo_vmware.api [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Task: {'id': task-1294093, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.935835] env[63021]: DEBUG oslo_concurrency.lockutils [req-0242a54b-518a-4493-81b0-49c262c7199d req-8c807ada-979c-4ea2-a477-e6fc073a0562 service nova] Releasing lock "refresh_cache-2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.003320] env[63021]: DEBUG nova.virt.hardware [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 967.003651] env[63021]: DEBUG nova.virt.hardware [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 967.003877] env[63021]: DEBUG nova.virt.hardware [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 967.007944] env[63021]: DEBUG nova.virt.hardware [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 967.007944] env[63021]: DEBUG nova.virt.hardware [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 967.007944] env[63021]: DEBUG nova.virt.hardware [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 967.007944] env[63021]: DEBUG nova.virt.hardware [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 967.007944] env[63021]: DEBUG nova.virt.hardware [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 967.007944] env[63021]: DEBUG nova.virt.hardware [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 967.007944] env[63021]: DEBUG nova.virt.hardware [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 967.007944] env[63021]: DEBUG nova.virt.hardware [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 967.011727] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Reconfiguring VM instance instance-00000055 to detach disk 2000 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 967.012137] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8beb8d7f-0a7c-42b4-b451-dae9c31a0a9f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.037039] env[63021]: DEBUG oslo_vmware.api [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 967.037039] env[63021]: value = "task-1294094" [ 967.037039] env[63021]: _type = "Task" [ 967.037039] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.051824] env[63021]: DEBUG oslo_vmware.api [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294094, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.056726] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Releasing lock "refresh_cache-cee7a86d-91e2-4b57-b6cd-29056da6a414" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.057018] env[63021]: DEBUG nova.compute.manager [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Instance network_info: |[{"id": "7aec4a00-7fef-4c0f-a168-2cfc26dbe0be", "address": "fa:16:3e:f3:38:47", "network": {"id": "81917a55-4e41-4898-9f4e-4fced52b5a86", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1793837736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e238d5c1668540d2a9fffd3fd832d9b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7aec4a00-7f", "ovs_interfaceid": "7aec4a00-7fef-4c0f-a168-2cfc26dbe0be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 967.057550] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f3:38:47', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '535b175f-71d3-4226-81ab-ca253f27fedd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7aec4a00-7fef-4c0f-a168-2cfc26dbe0be', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 967.066964] env[63021]: DEBUG oslo.service.loopingcall [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 967.066964] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 967.066964] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3b9c27ac-ab67-4112-abac-592fa8fda170 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.093377] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 967.093377] env[63021]: value = "task-1294095" [ 967.093377] env[63021]: _type = "Task" [ 967.093377] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.107561] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294095, 'name': CreateVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.108147] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3c8abbcd-9587-4f83-8d94-d536160299e4 tempest-ServersV294TestFqdnHostnames-426232371 tempest-ServersV294TestFqdnHostnames-426232371-project-member] Lock "ed26dff1-fed0-4baf-ad41-d14850254aec" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.257s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.171565] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1294092, 'name': CopyVirtualDisk_Task} progress is 29%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.207170] env[63021]: INFO nova.compute.manager [-] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Took 1.27 seconds to deallocate network for instance. [ 967.271898] env[63021]: DEBUG nova.compute.manager [req-33af0ffb-90d4-4916-a4e7-8174704ece72 req-bd9fdcd6-a9b0-46ea-9889-7165d6769933 service nova] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Received event network-vif-plugged-7aec4a00-7fef-4c0f-a168-2cfc26dbe0be {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 967.272276] env[63021]: DEBUG oslo_concurrency.lockutils [req-33af0ffb-90d4-4916-a4e7-8174704ece72 req-bd9fdcd6-a9b0-46ea-9889-7165d6769933 service nova] Acquiring lock "cee7a86d-91e2-4b57-b6cd-29056da6a414-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.273010] env[63021]: DEBUG oslo_concurrency.lockutils [req-33af0ffb-90d4-4916-a4e7-8174704ece72 req-bd9fdcd6-a9b0-46ea-9889-7165d6769933 service nova] Lock "cee7a86d-91e2-4b57-b6cd-29056da6a414-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.273308] env[63021]: DEBUG oslo_concurrency.lockutils [req-33af0ffb-90d4-4916-a4e7-8174704ece72 req-bd9fdcd6-a9b0-46ea-9889-7165d6769933 service nova] Lock "cee7a86d-91e2-4b57-b6cd-29056da6a414-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.273565] env[63021]: DEBUG nova.compute.manager [req-33af0ffb-90d4-4916-a4e7-8174704ece72 req-bd9fdcd6-a9b0-46ea-9889-7165d6769933 service nova] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] No waiting events found dispatching network-vif-plugged-7aec4a00-7fef-4c0f-a168-2cfc26dbe0be {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 967.273838] env[63021]: WARNING nova.compute.manager [req-33af0ffb-90d4-4916-a4e7-8174704ece72 req-bd9fdcd6-a9b0-46ea-9889-7165d6769933 service nova] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Received unexpected event network-vif-plugged-7aec4a00-7fef-4c0f-a168-2cfc26dbe0be for instance with vm_state building and task_state spawning. [ 967.273941] env[63021]: DEBUG nova.compute.manager [req-33af0ffb-90d4-4916-a4e7-8174704ece72 req-bd9fdcd6-a9b0-46ea-9889-7165d6769933 service nova] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Received event network-changed-7aec4a00-7fef-4c0f-a168-2cfc26dbe0be {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 967.274112] env[63021]: DEBUG nova.compute.manager [req-33af0ffb-90d4-4916-a4e7-8174704ece72 req-bd9fdcd6-a9b0-46ea-9889-7165d6769933 service nova] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Refreshing instance network info cache due to event network-changed-7aec4a00-7fef-4c0f-a168-2cfc26dbe0be. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 967.274313] env[63021]: DEBUG oslo_concurrency.lockutils [req-33af0ffb-90d4-4916-a4e7-8174704ece72 req-bd9fdcd6-a9b0-46ea-9889-7165d6769933 service nova] Acquiring lock "refresh_cache-cee7a86d-91e2-4b57-b6cd-29056da6a414" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.274457] env[63021]: DEBUG oslo_concurrency.lockutils [req-33af0ffb-90d4-4916-a4e7-8174704ece72 req-bd9fdcd6-a9b0-46ea-9889-7165d6769933 service nova] Acquired lock "refresh_cache-cee7a86d-91e2-4b57-b6cd-29056da6a414" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.274647] env[63021]: DEBUG nova.network.neutron [req-33af0ffb-90d4-4916-a4e7-8174704ece72 req-bd9fdcd6-a9b0-46ea-9889-7165d6769933 service nova] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Refreshing network info cache for port 7aec4a00-7fef-4c0f-a168-2cfc26dbe0be {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 967.402355] env[63021]: DEBUG oslo_vmware.api [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Task: {'id': task-1294093, 'name': PowerOnVM_Task, 'duration_secs': 0.971624} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.402739] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 967.403054] env[63021]: INFO nova.compute.manager [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Took 8.95 seconds to spawn the instance on the hypervisor. [ 967.403464] env[63021]: DEBUG nova.compute.manager [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 967.404510] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-769844bc-fa66-452e-a0cf-fadc4339e94c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.552340] env[63021]: DEBUG oslo_vmware.api [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294094, 'name': ReconfigVM_Task, 'duration_secs': 0.405278} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.552700] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Reconfigured VM instance instance-00000055 to detach disk 2000 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 967.554269] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-870907c3-8f86-4f75-bf28-04f0d44501f5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.584887] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] 207c8197-6191-4837-8ead-8c24aa5b35bc/207c8197-6191-4837-8ead-8c24aa5b35bc.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 967.590723] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-96322875-4ddc-45ad-a17f-4132d05c1146 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.617414] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294095, 'name': CreateVM_Task, 'duration_secs': 0.511294} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.618201] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 967.619056] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.619247] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.619603] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 967.620681] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-198a528f-bb80-4733-95eb-0705869772c7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.622677] env[63021]: DEBUG oslo_vmware.api [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 967.622677] env[63021]: value = "task-1294096" [ 967.622677] env[63021]: _type = "Task" [ 967.622677] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.632310] env[63021]: DEBUG oslo_vmware.api [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 967.632310] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52fcc000-d7b4-6fc8-b9fb-6e3daec3c12f" [ 967.632310] env[63021]: _type = "Task" [ 967.632310] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.641899] env[63021]: DEBUG oslo_vmware.api [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294096, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.651840] env[63021]: DEBUG oslo_vmware.api [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52fcc000-d7b4-6fc8-b9fb-6e3daec3c12f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.671501] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1294092, 'name': CopyVirtualDisk_Task} progress is 49%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.715754] env[63021]: DEBUG oslo_concurrency.lockutils [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.821545] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc7de034-db1b-4b05-b502-5a6a59615bdc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.835678] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2185410b-5a83-4af2-8b45-97738dfa73a7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.874551] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee200df5-6ac7-4f8d-a2e8-27b5f3e24433 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.886578] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c97d12e1-ed6f-4419-852b-c196089cd265 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.908029] env[63021]: DEBUG nova.compute.provider_tree [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 967.929382] env[63021]: INFO nova.compute.manager [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Took 18.77 seconds to build instance. [ 968.140634] env[63021]: DEBUG oslo_vmware.api [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294096, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.147040] env[63021]: DEBUG oslo_vmware.api [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52fcc000-d7b4-6fc8-b9fb-6e3daec3c12f, 'name': SearchDatastore_Task, 'duration_secs': 0.029928} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.147451] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.147698] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 968.148019] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.148217] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.148414] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 968.148705] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b95b396d-377b-4db2-ae30-92ee21331d2d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.163404] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 968.163657] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 968.164936] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b9a89a01-383a-4369-8f57-7aa0ff52404b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.173077] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1294092, 'name': CopyVirtualDisk_Task} progress is 71%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.174896] env[63021]: DEBUG oslo_vmware.api [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 968.174896] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5228bd7c-99c2-3797-19bb-9bfef54450b1" [ 968.174896] env[63021]: _type = "Task" [ 968.174896] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.183795] env[63021]: DEBUG oslo_vmware.api [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5228bd7c-99c2-3797-19bb-9bfef54450b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.377625] env[63021]: DEBUG nova.network.neutron [req-33af0ffb-90d4-4916-a4e7-8174704ece72 req-bd9fdcd6-a9b0-46ea-9889-7165d6769933 service nova] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Updated VIF entry in instance network info cache for port 7aec4a00-7fef-4c0f-a168-2cfc26dbe0be. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 968.378055] env[63021]: DEBUG nova.network.neutron [req-33af0ffb-90d4-4916-a4e7-8174704ece72 req-bd9fdcd6-a9b0-46ea-9889-7165d6769933 service nova] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Updating instance_info_cache with network_info: [{"id": "7aec4a00-7fef-4c0f-a168-2cfc26dbe0be", "address": "fa:16:3e:f3:38:47", "network": {"id": "81917a55-4e41-4898-9f4e-4fced52b5a86", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1793837736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e238d5c1668540d2a9fffd3fd832d9b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7aec4a00-7f", "ovs_interfaceid": "7aec4a00-7fef-4c0f-a168-2cfc26dbe0be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.409012] env[63021]: DEBUG nova.scheduler.client.report [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 968.431886] env[63021]: DEBUG oslo_concurrency.lockutils [None req-941456d0-4781-4ba9-add0-bebab32e2e02 tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Lock "7df37eb7-ffc1-4b8e-9aba-caaf3c5338df" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.282s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.634331] env[63021]: DEBUG oslo_vmware.api [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294096, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.666584] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1294092, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.685696] env[63021]: DEBUG oslo_vmware.api [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5228bd7c-99c2-3797-19bb-9bfef54450b1, 'name': SearchDatastore_Task, 'duration_secs': 0.02514} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.686689] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-680bb9a7-1826-47f7-8a77-468d1aaff650 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.693083] env[63021]: DEBUG oslo_vmware.api [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 968.693083] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]526bdbb5-586c-8e4a-780d-27d39aaa1b78" [ 968.693083] env[63021]: _type = "Task" [ 968.693083] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.702831] env[63021]: DEBUG oslo_vmware.api [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]526bdbb5-586c-8e4a-780d-27d39aaa1b78, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.884011] env[63021]: DEBUG oslo_concurrency.lockutils [req-33af0ffb-90d4-4916-a4e7-8174704ece72 req-bd9fdcd6-a9b0-46ea-9889-7165d6769933 service nova] Releasing lock "refresh_cache-cee7a86d-91e2-4b57-b6cd-29056da6a414" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.884215] env[63021]: DEBUG nova.compute.manager [req-33af0ffb-90d4-4916-a4e7-8174704ece72 req-bd9fdcd6-a9b0-46ea-9889-7165d6769933 service nova] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Received event network-vif-deleted-bb825b85-647f-45b0-b351-ec7921e3ae5e {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 968.915501] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.346s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.916240] env[63021]: DEBUG nova.compute.manager [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 968.922182] env[63021]: DEBUG oslo_concurrency.lockutils [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.206s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.922454] env[63021]: DEBUG nova.objects.instance [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lazy-loading 'resources' on Instance uuid 919887c8-171d-48fe-8c6b-2abf5f55d6ab {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 969.134338] env[63021]: DEBUG oslo_vmware.api [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294096, 'name': ReconfigVM_Task, 'duration_secs': 1.364898} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.134887] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Reconfigured VM instance instance-00000055 to attach disk [datastore2] 207c8197-6191-4837-8ead-8c24aa5b35bc/207c8197-6191-4837-8ead-8c24aa5b35bc.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 969.135213] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Updating instance '207c8197-6191-4837-8ead-8c24aa5b35bc' progress to 50 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 969.167449] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1294092, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.734784} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.167774] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d9df5b42-2840-4afc-82fa-268aed408575/d9df5b42-2840-4afc-82fa-268aed408575.vmdk to [datastore2] 45c9e2be-eb19-4b83-b280-c9eeaddfccba/45c9e2be-eb19-4b83-b280-c9eeaddfccba.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 969.168584] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23d300a6-37d5-42d3-b7d8-3ab275ee7341 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.191757] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] 45c9e2be-eb19-4b83-b280-c9eeaddfccba/45c9e2be-eb19-4b83-b280-c9eeaddfccba.vmdk or device None with type streamOptimized {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 969.192895] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6da63e49-cb89-443f-a988-6063f28d34c1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.218377] env[63021]: DEBUG oslo_vmware.api [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]526bdbb5-586c-8e4a-780d-27d39aaa1b78, 'name': SearchDatastore_Task, 'duration_secs': 0.020384} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.219765] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.220065] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] cee7a86d-91e2-4b57-b6cd-29056da6a414/cee7a86d-91e2-4b57-b6cd-29056da6a414.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 969.220445] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Waiting for the task: (returnval){ [ 969.220445] env[63021]: value = "task-1294097" [ 969.220445] env[63021]: _type = "Task" [ 969.220445] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.220680] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-44b7b2bd-032d-4be9-9307-215a4e051760 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.233955] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1294097, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.235519] env[63021]: DEBUG oslo_vmware.api [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 969.235519] env[63021]: value = "task-1294098" [ 969.235519] env[63021]: _type = "Task" [ 969.235519] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.243611] env[63021]: DEBUG oslo_vmware.api [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294098, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.425522] env[63021]: DEBUG nova.compute.utils [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 969.429900] env[63021]: DEBUG nova.compute.manager [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 969.430102] env[63021]: DEBUG nova.network.neutron [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 969.480931] env[63021]: DEBUG nova.policy [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6f9dc7ceda4f4ddbba1de296ee4cc814', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b672b1d7d7ad454ebfa3e2fd2421c754', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 969.624996] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3a0718f-ef3c-49cc-9718-5b66d9d57d0d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.635696] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9deb3f06-4d0b-4715-ac13-ab4509a148f1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.646063] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b650e96d-1ba5-4546-a1ce-a000ec0b9a23 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.691996] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3076835-d70d-4be9-9741-fa6d4db81a74 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.695484] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bc09f51-2ae7-4df8-af98-cc53315af1fb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.719904] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fcac7db-3592-4e2e-9d80-0350a9bf0dad {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.724407] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Updating instance '207c8197-6191-4837-8ead-8c24aa5b35bc' progress to 67 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 969.741633] env[63021]: DEBUG nova.compute.provider_tree [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 969.749389] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1294097, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.755080] env[63021]: DEBUG oslo_vmware.api [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294098, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.785572] env[63021]: DEBUG nova.network.neutron [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Successfully created port: db333f88-b16f-439c-99fe-ca5884a09dc1 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 969.936148] env[63021]: DEBUG nova.compute.manager [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 970.251016] env[63021]: DEBUG nova.scheduler.client.report [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 970.257072] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1294097, 'name': ReconfigVM_Task, 'duration_secs': 0.903425} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.258017] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Reconfigured VM instance instance-00000044 to attach disk [datastore2] 45c9e2be-eb19-4b83-b280-c9eeaddfccba/45c9e2be-eb19-4b83-b280-c9eeaddfccba.vmdk or device None with type streamOptimized {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 970.264926] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f1deabe8-fe8a-4808-9a82-ff11dd4f4b1e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.267219] env[63021]: DEBUG oslo_vmware.api [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294098, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.275490] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Waiting for the task: (returnval){ [ 970.275490] env[63021]: value = "task-1294099" [ 970.275490] env[63021]: _type = "Task" [ 970.275490] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.287882] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1294099, 'name': Rename_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.298254] env[63021]: DEBUG nova.network.neutron [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Port a7ecacae-44cb-4b30-892f-d0f5ef3ad9d5 binding to destination host cpu-1 is already ACTIVE {{(pid=63021) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 970.328802] env[63021]: DEBUG oslo_concurrency.lockutils [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Acquiring lock "7df37eb7-ffc1-4b8e-9aba-caaf3c5338df" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.329130] env[63021]: DEBUG oslo_concurrency.lockutils [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Lock "7df37eb7-ffc1-4b8e-9aba-caaf3c5338df" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.329352] env[63021]: DEBUG oslo_concurrency.lockutils [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Acquiring lock "7df37eb7-ffc1-4b8e-9aba-caaf3c5338df-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.329551] env[63021]: DEBUG oslo_concurrency.lockutils [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Lock "7df37eb7-ffc1-4b8e-9aba-caaf3c5338df-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.329749] env[63021]: DEBUG oslo_concurrency.lockutils [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Lock "7df37eb7-ffc1-4b8e-9aba-caaf3c5338df-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.332366] env[63021]: INFO nova.compute.manager [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Terminating instance [ 970.335191] env[63021]: DEBUG nova.compute.manager [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 970.335632] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 970.337133] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d367fa34-e7e2-4eb8-a84a-054693ce2aee {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.346458] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 970.347124] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c44396c0-c10c-4195-af94-537de3bb57ec {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.353708] env[63021]: DEBUG oslo_vmware.api [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Waiting for the task: (returnval){ [ 970.353708] env[63021]: value = "task-1294100" [ 970.353708] env[63021]: _type = "Task" [ 970.353708] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.363482] env[63021]: DEBUG oslo_vmware.api [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Task: {'id': task-1294100, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.753475] env[63021]: DEBUG oslo_vmware.api [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294098, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.137303} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.753745] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] cee7a86d-91e2-4b57-b6cd-29056da6a414/cee7a86d-91e2-4b57-b6cd-29056da6a414.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 970.753970] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 970.754265] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1b55ea8c-4153-42d7-af50-c32c6ba0df92 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.756603] env[63021]: DEBUG oslo_concurrency.lockutils [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.835s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.763858] env[63021]: DEBUG oslo_vmware.api [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 970.763858] env[63021]: value = "task-1294101" [ 970.763858] env[63021]: _type = "Task" [ 970.763858] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.772845] env[63021]: DEBUG oslo_vmware.api [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294101, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.781064] env[63021]: INFO nova.scheduler.client.report [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Deleted allocations for instance 919887c8-171d-48fe-8c6b-2abf5f55d6ab [ 970.787760] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1294099, 'name': Rename_Task, 'duration_secs': 0.288308} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.788228] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 970.788714] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6041c2e4-6a90-4601-b0ed-e274aae61732 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.794109] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Waiting for the task: (returnval){ [ 970.794109] env[63021]: value = "task-1294102" [ 970.794109] env[63021]: _type = "Task" [ 970.794109] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.804025] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1294102, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.865207] env[63021]: DEBUG oslo_vmware.api [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Task: {'id': task-1294100, 'name': PowerOffVM_Task, 'duration_secs': 0.191043} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.865519] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 970.866430] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 970.866430] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1c30d8e0-3ebb-45a4-b8ee-bca5cb094334 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.927231] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 970.927530] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Deleting contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 970.927726] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Deleting the datastore file [datastore2] 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 970.928369] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-37d8a846-d8e1-491d-830b-54f78c75e6c6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.934389] env[63021]: DEBUG oslo_vmware.api [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Waiting for the task: (returnval){ [ 970.934389] env[63021]: value = "task-1294104" [ 970.934389] env[63021]: _type = "Task" [ 970.934389] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.942833] env[63021]: DEBUG oslo_vmware.api [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Task: {'id': task-1294104, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.949150] env[63021]: DEBUG nova.compute.manager [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 970.977845] env[63021]: DEBUG nova.virt.hardware [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 970.978156] env[63021]: DEBUG nova.virt.hardware [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 970.978365] env[63021]: DEBUG nova.virt.hardware [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 970.978570] env[63021]: DEBUG nova.virt.hardware [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 970.978720] env[63021]: DEBUG nova.virt.hardware [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 970.978869] env[63021]: DEBUG nova.virt.hardware [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 970.979101] env[63021]: DEBUG nova.virt.hardware [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 970.979271] env[63021]: DEBUG nova.virt.hardware [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 970.979442] env[63021]: DEBUG nova.virt.hardware [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 970.979606] env[63021]: DEBUG nova.virt.hardware [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 970.979991] env[63021]: DEBUG nova.virt.hardware [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 970.980868] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-458a02a5-d9b7-4929-af3e-dd437bd1f18a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.988783] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-227daf31-d73a-49f9-a005-dc2ebcc2f691 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.273967] env[63021]: DEBUG oslo_vmware.api [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294101, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.42367} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.274386] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 971.275217] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed7a3623-efe6-4716-b4c3-f03c2d78f346 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.301179] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] cee7a86d-91e2-4b57-b6cd-29056da6a414/cee7a86d-91e2-4b57-b6cd-29056da6a414.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 971.301946] env[63021]: DEBUG oslo_concurrency.lockutils [None req-61c866ad-4751-476e-bd96-35704cf1a785 tempest-ImagesTestJSON-275894372 tempest-ImagesTestJSON-275894372-project-member] Lock "919887c8-171d-48fe-8c6b-2abf5f55d6ab" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.077s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.302918] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e6727495-5a50-497b-a696-7bad6b4baaef {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.346203] env[63021]: DEBUG oslo_concurrency.lockutils [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "207c8197-6191-4837-8ead-8c24aa5b35bc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.346203] env[63021]: DEBUG oslo_concurrency.lockutils [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "207c8197-6191-4837-8ead-8c24aa5b35bc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.346313] env[63021]: DEBUG oslo_concurrency.lockutils [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "207c8197-6191-4837-8ead-8c24aa5b35bc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.347681] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1294102, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.347920] env[63021]: DEBUG oslo_vmware.api [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 971.347920] env[63021]: value = "task-1294105" [ 971.347920] env[63021]: _type = "Task" [ 971.347920] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.352722] env[63021]: DEBUG nova.network.neutron [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Successfully updated port: db333f88-b16f-439c-99fe-ca5884a09dc1 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 971.362023] env[63021]: DEBUG oslo_vmware.api [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294105, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.374529] env[63021]: DEBUG nova.compute.manager [req-633adc71-b3cd-49ae-a15d-7668aed4df51 req-9f188432-326c-44cd-8ced-e29a2c2a6981 service nova] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Received event network-vif-plugged-db333f88-b16f-439c-99fe-ca5884a09dc1 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 971.374749] env[63021]: DEBUG oslo_concurrency.lockutils [req-633adc71-b3cd-49ae-a15d-7668aed4df51 req-9f188432-326c-44cd-8ced-e29a2c2a6981 service nova] Acquiring lock "67f736d9-82ec-448e-8b72-84fdd704aa8d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.374962] env[63021]: DEBUG oslo_concurrency.lockutils [req-633adc71-b3cd-49ae-a15d-7668aed4df51 req-9f188432-326c-44cd-8ced-e29a2c2a6981 service nova] Lock "67f736d9-82ec-448e-8b72-84fdd704aa8d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.375151] env[63021]: DEBUG oslo_concurrency.lockutils [req-633adc71-b3cd-49ae-a15d-7668aed4df51 req-9f188432-326c-44cd-8ced-e29a2c2a6981 service nova] Lock "67f736d9-82ec-448e-8b72-84fdd704aa8d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.375319] env[63021]: DEBUG nova.compute.manager [req-633adc71-b3cd-49ae-a15d-7668aed4df51 req-9f188432-326c-44cd-8ced-e29a2c2a6981 service nova] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] No waiting events found dispatching network-vif-plugged-db333f88-b16f-439c-99fe-ca5884a09dc1 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 971.375489] env[63021]: WARNING nova.compute.manager [req-633adc71-b3cd-49ae-a15d-7668aed4df51 req-9f188432-326c-44cd-8ced-e29a2c2a6981 service nova] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Received unexpected event network-vif-plugged-db333f88-b16f-439c-99fe-ca5884a09dc1 for instance with vm_state building and task_state spawning. [ 971.445603] env[63021]: DEBUG oslo_vmware.api [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Task: {'id': task-1294104, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.358788} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.445791] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 971.446259] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Deleted contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 971.446259] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 971.446468] env[63021]: INFO nova.compute.manager [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Took 1.11 seconds to destroy the instance on the hypervisor. [ 971.446703] env[63021]: DEBUG oslo.service.loopingcall [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 971.446922] env[63021]: DEBUG nova.compute.manager [-] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 971.447030] env[63021]: DEBUG nova.network.neutron [-] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 971.813360] env[63021]: DEBUG oslo_vmware.api [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1294102, 'name': PowerOnVM_Task, 'duration_secs': 0.714909} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.815207] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 971.864034] env[63021]: DEBUG oslo_vmware.api [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294105, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.865560] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "refresh_cache-67f736d9-82ec-448e-8b72-84fdd704aa8d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.865560] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquired lock "refresh_cache-67f736d9-82ec-448e-8b72-84fdd704aa8d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.865560] env[63021]: DEBUG nova.network.neutron [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 971.924654] env[63021]: DEBUG nova.compute.manager [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 971.927019] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f087bc5-e857-4a2c-bbc7-f224426d0c6c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.194204] env[63021]: DEBUG nova.network.neutron [-] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.365159] env[63021]: DEBUG oslo_vmware.api [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294105, 'name': ReconfigVM_Task, 'duration_secs': 0.579966} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.365477] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Reconfigured VM instance instance-0000005a to attach disk [datastore1] cee7a86d-91e2-4b57-b6cd-29056da6a414/cee7a86d-91e2-4b57-b6cd-29056da6a414.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 972.366515] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7285b0a4-9a35-47b3-b23f-8a0602ded12b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.373756] env[63021]: DEBUG oslo_vmware.api [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 972.373756] env[63021]: value = "task-1294106" [ 972.373756] env[63021]: _type = "Task" [ 972.373756] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.382388] env[63021]: DEBUG oslo_vmware.api [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294106, 'name': Rename_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.387678] env[63021]: DEBUG oslo_concurrency.lockutils [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "refresh_cache-207c8197-6191-4837-8ead-8c24aa5b35bc" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.387872] env[63021]: DEBUG oslo_concurrency.lockutils [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquired lock "refresh_cache-207c8197-6191-4837-8ead-8c24aa5b35bc" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.388055] env[63021]: DEBUG nova.network.neutron [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 972.419327] env[63021]: DEBUG nova.network.neutron [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 972.443939] env[63021]: DEBUG oslo_concurrency.lockutils [None req-795e3952-63a3-4c7e-9970-3524ef4e8b14 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lock "45c9e2be-eb19-4b83-b280-c9eeaddfccba" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 31.587s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.678045] env[63021]: DEBUG nova.network.neutron [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Updating instance_info_cache with network_info: [{"id": "db333f88-b16f-439c-99fe-ca5884a09dc1", "address": "fa:16:3e:40:f8:6d", "network": {"id": "3aaeed35-f452-4144-a259-8f440f8a8bf4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-229941363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b672b1d7d7ad454ebfa3e2fd2421c754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb333f88-b1", "ovs_interfaceid": "db333f88-b16f-439c-99fe-ca5884a09dc1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.696661] env[63021]: INFO nova.compute.manager [-] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Took 1.25 seconds to deallocate network for instance. [ 972.884324] env[63021]: DEBUG oslo_vmware.api [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294106, 'name': Rename_Task, 'duration_secs': 0.153048} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.884662] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 972.884975] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-624b4836-78e9-4e48-8b35-888091ea3ada {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.893154] env[63021]: DEBUG oslo_vmware.api [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 972.893154] env[63021]: value = "task-1294107" [ 972.893154] env[63021]: _type = "Task" [ 972.893154] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.900993] env[63021]: DEBUG oslo_vmware.api [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294107, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.101974] env[63021]: DEBUG nova.network.neutron [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Updating instance_info_cache with network_info: [{"id": "a7ecacae-44cb-4b30-892f-d0f5ef3ad9d5", "address": "fa:16:3e:b7:1c:44", "network": {"id": "708d0dc3-c651-4001-9638-1d3538ff76c9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-983471056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "805af233efa147b3982fb794cd9e6f73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7ecacae-44", "ovs_interfaceid": "a7ecacae-44cb-4b30-892f-d0f5ef3ad9d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.181923] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Releasing lock "refresh_cache-67f736d9-82ec-448e-8b72-84fdd704aa8d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.183023] env[63021]: DEBUG nova.compute.manager [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Instance network_info: |[{"id": "db333f88-b16f-439c-99fe-ca5884a09dc1", "address": "fa:16:3e:40:f8:6d", "network": {"id": "3aaeed35-f452-4144-a259-8f440f8a8bf4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-229941363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b672b1d7d7ad454ebfa3e2fd2421c754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb333f88-b1", "ovs_interfaceid": "db333f88-b16f-439c-99fe-ca5884a09dc1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 973.183023] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:40:f8:6d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ead20342-9afa-435e-a22b-b4a903457712', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'db333f88-b16f-439c-99fe-ca5884a09dc1', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 973.191608] env[63021]: DEBUG oslo.service.loopingcall [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 973.192136] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 973.192238] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7451156a-40b2-446c-b4c6-edf9596d08ac {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.209495] env[63021]: DEBUG oslo_concurrency.lockutils [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.209784] env[63021]: DEBUG oslo_concurrency.lockutils [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.210068] env[63021]: DEBUG nova.objects.instance [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Lazy-loading 'resources' on Instance uuid 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 973.217365] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 973.217365] env[63021]: value = "task-1294108" [ 973.217365] env[63021]: _type = "Task" [ 973.217365] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.226541] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294108, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.403899] env[63021]: DEBUG oslo_vmware.api [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294107, 'name': PowerOnVM_Task, 'duration_secs': 0.487891} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.404598] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 973.404928] env[63021]: INFO nova.compute.manager [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Took 8.47 seconds to spawn the instance on the hypervisor. [ 973.405196] env[63021]: DEBUG nova.compute.manager [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 973.406387] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e831c40-fd4c-441f-af80-c096ed6df12b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.532677] env[63021]: DEBUG nova.compute.manager [req-8dfb7f35-e1a0-4814-889a-05d8634659d3 req-b4bdfafa-54b0-4ab9-a72f-2b3b8e9d1aca service nova] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Received event network-changed-db333f88-b16f-439c-99fe-ca5884a09dc1 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 973.532880] env[63021]: DEBUG nova.compute.manager [req-8dfb7f35-e1a0-4814-889a-05d8634659d3 req-b4bdfafa-54b0-4ab9-a72f-2b3b8e9d1aca service nova] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Refreshing instance network info cache due to event network-changed-db333f88-b16f-439c-99fe-ca5884a09dc1. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 973.533148] env[63021]: DEBUG oslo_concurrency.lockutils [req-8dfb7f35-e1a0-4814-889a-05d8634659d3 req-b4bdfafa-54b0-4ab9-a72f-2b3b8e9d1aca service nova] Acquiring lock "refresh_cache-67f736d9-82ec-448e-8b72-84fdd704aa8d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.533251] env[63021]: DEBUG oslo_concurrency.lockutils [req-8dfb7f35-e1a0-4814-889a-05d8634659d3 req-b4bdfafa-54b0-4ab9-a72f-2b3b8e9d1aca service nova] Acquired lock "refresh_cache-67f736d9-82ec-448e-8b72-84fdd704aa8d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.533415] env[63021]: DEBUG nova.network.neutron [req-8dfb7f35-e1a0-4814-889a-05d8634659d3 req-b4bdfafa-54b0-4ab9-a72f-2b3b8e9d1aca service nova] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Refreshing network info cache for port db333f88-b16f-439c-99fe-ca5884a09dc1 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 973.605797] env[63021]: DEBUG oslo_concurrency.lockutils [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Releasing lock "refresh_cache-207c8197-6191-4837-8ead-8c24aa5b35bc" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.726912] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294108, 'name': CreateVM_Task} progress is 25%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.872900] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64e92c68-d562-4da6-b2d7-eea96fa4730a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.880407] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-900e5f6e-3bd4-4857-8d38-2b977fcb6e35 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.913336] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39682e9a-9773-4499-9428-75d7f8ac39f7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.924385] env[63021]: INFO nova.compute.manager [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Took 23.14 seconds to build instance. [ 973.926258] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b9ddc4b-0351-4209-a46d-506540b0a940 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.940744] env[63021]: DEBUG nova.compute.provider_tree [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 974.129518] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3d9649a-3e43-40c5-b7dc-d5dab615fcad {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.152773] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f6b2791-586e-458f-9e2b-cce6c60e6bde {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.160163] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Updating instance '207c8197-6191-4837-8ead-8c24aa5b35bc' progress to 83 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 974.229745] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294108, 'name': CreateVM_Task, 'duration_secs': 0.727555} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.229745] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 974.230362] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.230611] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.231030] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 974.231590] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ac813de-89ea-4394-8efd-fdf45be53bee {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.236268] env[63021]: DEBUG oslo_vmware.api [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 974.236268] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]527b2957-dbac-b297-959d-6a11a4a8de66" [ 974.236268] env[63021]: _type = "Task" [ 974.236268] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.244346] env[63021]: DEBUG oslo_vmware.api [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]527b2957-dbac-b297-959d-6a11a4a8de66, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.298169] env[63021]: DEBUG nova.network.neutron [req-8dfb7f35-e1a0-4814-889a-05d8634659d3 req-b4bdfafa-54b0-4ab9-a72f-2b3b8e9d1aca service nova] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Updated VIF entry in instance network info cache for port db333f88-b16f-439c-99fe-ca5884a09dc1. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 974.298618] env[63021]: DEBUG nova.network.neutron [req-8dfb7f35-e1a0-4814-889a-05d8634659d3 req-b4bdfafa-54b0-4ab9-a72f-2b3b8e9d1aca service nova] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Updating instance_info_cache with network_info: [{"id": "db333f88-b16f-439c-99fe-ca5884a09dc1", "address": "fa:16:3e:40:f8:6d", "network": {"id": "3aaeed35-f452-4144-a259-8f440f8a8bf4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-229941363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b672b1d7d7ad454ebfa3e2fd2421c754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb333f88-b1", "ovs_interfaceid": "db333f88-b16f-439c-99fe-ca5884a09dc1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.431377] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9f27d0f0-0704-4b92-a7b7-7f37d67b812d tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "cee7a86d-91e2-4b57-b6cd-29056da6a414" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.657s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.447187] env[63021]: DEBUG nova.scheduler.client.report [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 974.509203] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "cee7a86d-91e2-4b57-b6cd-29056da6a414" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.509596] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "cee7a86d-91e2-4b57-b6cd-29056da6a414" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.509864] env[63021]: INFO nova.compute.manager [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Shelving [ 974.666384] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 974.666698] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0c78685d-56d1-4b72-b09d-b36b5588ede5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.674681] env[63021]: DEBUG oslo_vmware.api [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 974.674681] env[63021]: value = "task-1294109" [ 974.674681] env[63021]: _type = "Task" [ 974.674681] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.685349] env[63021]: DEBUG oslo_vmware.api [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294109, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.746393] env[63021]: DEBUG oslo_vmware.api [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]527b2957-dbac-b297-959d-6a11a4a8de66, 'name': SearchDatastore_Task, 'duration_secs': 0.015952} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.746644] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.746874] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 974.747135] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.747306] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.747498] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 974.747804] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2df9d83a-13a2-40a2-90e5-df4da105d0c4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.764862] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 974.765064] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 974.765904] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b9e51e31-18bf-40c9-b3c5-14ca59f7789e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.770835] env[63021]: DEBUG oslo_vmware.api [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 974.770835] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5258588f-feed-c6a7-4948-96e7b1253baa" [ 974.770835] env[63021]: _type = "Task" [ 974.770835] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.778060] env[63021]: DEBUG oslo_vmware.api [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5258588f-feed-c6a7-4948-96e7b1253baa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.801209] env[63021]: DEBUG oslo_concurrency.lockutils [req-8dfb7f35-e1a0-4814-889a-05d8634659d3 req-b4bdfafa-54b0-4ab9-a72f-2b3b8e9d1aca service nova] Releasing lock "refresh_cache-67f736d9-82ec-448e-8b72-84fdd704aa8d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.801557] env[63021]: DEBUG nova.compute.manager [req-8dfb7f35-e1a0-4814-889a-05d8634659d3 req-b4bdfafa-54b0-4ab9-a72f-2b3b8e9d1aca service nova] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Received event network-vif-deleted-a8e61c39-aea7-48e6-9798-cbf508c50425 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 974.951954] env[63021]: DEBUG oslo_concurrency.lockutils [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.742s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.974565] env[63021]: INFO nova.scheduler.client.report [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Deleted allocations for instance 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df [ 975.018638] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 975.018867] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5849f8df-b111-4c95-a897-694154acb12a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.025779] env[63021]: DEBUG oslo_vmware.api [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 975.025779] env[63021]: value = "task-1294110" [ 975.025779] env[63021]: _type = "Task" [ 975.025779] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.034904] env[63021]: DEBUG oslo_vmware.api [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294110, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.186216] env[63021]: DEBUG oslo_vmware.api [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294109, 'name': PowerOnVM_Task, 'duration_secs': 0.41046} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.186623] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 975.186879] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-50842ef8-9335-4988-903c-9b3b6867158b tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Updating instance '207c8197-6191-4837-8ead-8c24aa5b35bc' progress to 100 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 975.281901] env[63021]: DEBUG oslo_vmware.api [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5258588f-feed-c6a7-4948-96e7b1253baa, 'name': SearchDatastore_Task, 'duration_secs': 0.052142} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.282726] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-006c2fe2-d621-45fa-b767-6b40fe21d10e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.289024] env[63021]: DEBUG oslo_vmware.api [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 975.289024] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52ef0deb-81c1-1375-2d98-189846647dfd" [ 975.289024] env[63021]: _type = "Task" [ 975.289024] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.295516] env[63021]: DEBUG oslo_vmware.api [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52ef0deb-81c1-1375-2d98-189846647dfd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.489668] env[63021]: DEBUG oslo_concurrency.lockutils [None req-11bb8557-c2f7-4ad1-9187-e3cc2c61166a tempest-ServerTagsTestJSON-800965522 tempest-ServerTagsTestJSON-800965522-project-member] Lock "7df37eb7-ffc1-4b8e-9aba-caaf3c5338df" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.160s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.536138] env[63021]: DEBUG oslo_vmware.api [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294110, 'name': PowerOffVM_Task, 'duration_secs': 0.256053} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.536402] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 975.537200] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae6873d5-054b-4942-a591-a11f43b6c066 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.556782] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d0697ee-e8ed-438f-bfdf-3f59e94aa659 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.797965] env[63021]: DEBUG oslo_vmware.api [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52ef0deb-81c1-1375-2d98-189846647dfd, 'name': SearchDatastore_Task, 'duration_secs': 0.011685} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.798214] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.798534] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 67f736d9-82ec-448e-8b72-84fdd704aa8d/67f736d9-82ec-448e-8b72-84fdd704aa8d.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 975.798805] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-272c4210-22d1-4176-afae-68f805733984 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.805298] env[63021]: DEBUG oslo_vmware.api [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 975.805298] env[63021]: value = "task-1294111" [ 975.805298] env[63021]: _type = "Task" [ 975.805298] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.812371] env[63021]: DEBUG oslo_vmware.api [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294111, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.067566] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Creating Snapshot of the VM instance {{(pid=63021) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 976.067937] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-91e66898-565b-4cd6-bd42-4aede5377e8d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.075735] env[63021]: DEBUG oslo_vmware.api [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 976.075735] env[63021]: value = "task-1294112" [ 976.075735] env[63021]: _type = "Task" [ 976.075735] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.087852] env[63021]: DEBUG oslo_vmware.api [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294112, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.318226] env[63021]: DEBUG oslo_vmware.api [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294111, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.589123] env[63021]: DEBUG oslo_vmware.api [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294112, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.817627] env[63021]: DEBUG oslo_vmware.api [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294111, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.726194} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.817896] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 67f736d9-82ec-448e-8b72-84fdd704aa8d/67f736d9-82ec-448e-8b72-84fdd704aa8d.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 976.818584] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 976.818841] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c15feae8-05ca-47a2-bed7-65283566da2d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.825847] env[63021]: DEBUG oslo_vmware.api [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 976.825847] env[63021]: value = "task-1294113" [ 976.825847] env[63021]: _type = "Task" [ 976.825847] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.836062] env[63021]: DEBUG oslo_vmware.api [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294113, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.086828] env[63021]: DEBUG oslo_vmware.api [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294112, 'name': CreateSnapshot_Task, 'duration_secs': 0.633801} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.087275] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Created Snapshot of the VM instance {{(pid=63021) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 977.087975] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8602f51d-a61d-4efe-b494-92cd3c9859a2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.256950] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 977.257225] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 977.335449] env[63021]: DEBUG oslo_vmware.api [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294113, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.207968} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.335728] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 977.336539] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b6f57fb-9735-4326-b370-ea7553729e29 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.358474] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] 67f736d9-82ec-448e-8b72-84fdd704aa8d/67f736d9-82ec-448e-8b72-84fdd704aa8d.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 977.358782] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-67045295-a69a-424e-b530-6447963fea3f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.379021] env[63021]: DEBUG oslo_vmware.api [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 977.379021] env[63021]: value = "task-1294114" [ 977.379021] env[63021]: _type = "Task" [ 977.379021] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.386892] env[63021]: DEBUG oslo_vmware.api [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294114, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.435477] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fe550412-75b7-4078-8f53-e9fe3264ec95 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "207c8197-6191-4837-8ead-8c24aa5b35bc" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.435860] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fe550412-75b7-4078-8f53-e9fe3264ec95 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "207c8197-6191-4837-8ead-8c24aa5b35bc" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.436181] env[63021]: DEBUG nova.compute.manager [None req-fe550412-75b7-4078-8f53-e9fe3264ec95 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Going to confirm migration 2 {{(pid=63021) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 977.606711] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Creating linked-clone VM from snapshot {{(pid=63021) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 977.607079] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-b63cec9d-42d6-41a5-9e6a-4fcd5b2323f7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.616059] env[63021]: DEBUG oslo_vmware.api [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 977.616059] env[63021]: value = "task-1294115" [ 977.616059] env[63021]: _type = "Task" [ 977.616059] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.624263] env[63021]: DEBUG oslo_vmware.api [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294115, 'name': CloneVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.764369] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 977.764524] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Starting heal instance info cache {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 977.890452] env[63021]: DEBUG oslo_vmware.api [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294114, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.998177] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fe550412-75b7-4078-8f53-e9fe3264ec95 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "refresh_cache-207c8197-6191-4837-8ead-8c24aa5b35bc" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.998372] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fe550412-75b7-4078-8f53-e9fe3264ec95 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquired lock "refresh_cache-207c8197-6191-4837-8ead-8c24aa5b35bc" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.998594] env[63021]: DEBUG nova.network.neutron [None req-fe550412-75b7-4078-8f53-e9fe3264ec95 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 977.998800] env[63021]: DEBUG nova.objects.instance [None req-fe550412-75b7-4078-8f53-e9fe3264ec95 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lazy-loading 'info_cache' on Instance uuid 207c8197-6191-4837-8ead-8c24aa5b35bc {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 978.126928] env[63021]: DEBUG oslo_vmware.api [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294115, 'name': CloneVM_Task} progress is 94%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.389830] env[63021]: DEBUG oslo_vmware.api [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294114, 'name': ReconfigVM_Task, 'duration_secs': 0.679756} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.390175] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Reconfigured VM instance instance-0000005b to attach disk [datastore1] 67f736d9-82ec-448e-8b72-84fdd704aa8d/67f736d9-82ec-448e-8b72-84fdd704aa8d.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 978.390820] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-84c8b12e-97e8-429a-9993-beb2958cb11a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.397429] env[63021]: DEBUG oslo_vmware.api [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 978.397429] env[63021]: value = "task-1294116" [ 978.397429] env[63021]: _type = "Task" [ 978.397429] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.405430] env[63021]: DEBUG oslo_vmware.api [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294116, 'name': Rename_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.629624] env[63021]: DEBUG oslo_vmware.api [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294115, 'name': CloneVM_Task} progress is 94%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.907087] env[63021]: DEBUG oslo_vmware.api [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294116, 'name': Rename_Task, 'duration_secs': 0.169286} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.907421] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 978.907681] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-36233d16-a988-4010-9415-12e09e20ae58 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.914171] env[63021]: DEBUG oslo_vmware.api [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 978.914171] env[63021]: value = "task-1294117" [ 978.914171] env[63021]: _type = "Task" [ 978.914171] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.925126] env[63021]: DEBUG oslo_vmware.api [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294117, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.130041] env[63021]: DEBUG oslo_vmware.api [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294115, 'name': CloneVM_Task, 'duration_secs': 1.41067} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.130354] env[63021]: INFO nova.virt.vmwareapi.vmops [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Created linked-clone VM from snapshot [ 979.131304] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-204273f2-04ff-4a77-a3b3-191be50962bf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.144023] env[63021]: DEBUG nova.virt.vmwareapi.images [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Uploading image 9f7298ed-bbc9-476d-b57e-755a41852bfe {{(pid=63021) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 979.177374] env[63021]: DEBUG oslo_vmware.rw_handles [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 979.177374] env[63021]: value = "vm-277573" [ 979.177374] env[63021]: _type = "VirtualMachine" [ 979.177374] env[63021]: }. {{(pid=63021) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 979.178414] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-69284476-0d88-4c28-8057-c070f45bac0a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.185367] env[63021]: DEBUG oslo_vmware.rw_handles [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lease: (returnval){ [ 979.185367] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52d63ddb-ac93-74d2-a333-09dba6321dcb" [ 979.185367] env[63021]: _type = "HttpNfcLease" [ 979.185367] env[63021]: } obtained for exporting VM: (result){ [ 979.185367] env[63021]: value = "vm-277573" [ 979.185367] env[63021]: _type = "VirtualMachine" [ 979.185367] env[63021]: }. {{(pid=63021) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 979.185914] env[63021]: DEBUG oslo_vmware.api [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the lease: (returnval){ [ 979.185914] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52d63ddb-ac93-74d2-a333-09dba6321dcb" [ 979.185914] env[63021]: _type = "HttpNfcLease" [ 979.185914] env[63021]: } to be ready. {{(pid=63021) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 979.201865] env[63021]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 979.201865] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52d63ddb-ac93-74d2-a333-09dba6321dcb" [ 979.201865] env[63021]: _type = "HttpNfcLease" [ 979.201865] env[63021]: } is ready. {{(pid=63021) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 979.202201] env[63021]: DEBUG oslo_vmware.rw_handles [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 979.202201] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52d63ddb-ac93-74d2-a333-09dba6321dcb" [ 979.202201] env[63021]: _type = "HttpNfcLease" [ 979.202201] env[63021]: }. {{(pid=63021) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 979.203144] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74023332-6a10-4e40-af37-33f6ffe408b4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.213020] env[63021]: DEBUG oslo_vmware.rw_handles [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526e9ee5-d828-90b0-5e90-db179d7100a0/disk-0.vmdk from lease info. {{(pid=63021) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 979.213020] env[63021]: DEBUG oslo_vmware.rw_handles [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526e9ee5-d828-90b0-5e90-db179d7100a0/disk-0.vmdk for reading. {{(pid=63021) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 979.311388] env[63021]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-c3d3f2ea-63fc-489e-879b-4b4622b9c59e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.425961] env[63021]: DEBUG oslo_vmware.api [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294117, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.436919] env[63021]: DEBUG nova.network.neutron [None req-fe550412-75b7-4078-8f53-e9fe3264ec95 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Updating instance_info_cache with network_info: [{"id": "a7ecacae-44cb-4b30-892f-d0f5ef3ad9d5", "address": "fa:16:3e:b7:1c:44", "network": {"id": "708d0dc3-c651-4001-9638-1d3538ff76c9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-983471056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "805af233efa147b3982fb794cd9e6f73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7ecacae-44", "ovs_interfaceid": "a7ecacae-44cb-4b30-892f-d0f5ef3ad9d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.927073] env[63021]: DEBUG oslo_vmware.api [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294117, 'name': PowerOnVM_Task, 'duration_secs': 0.745131} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.929732] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 979.929732] env[63021]: INFO nova.compute.manager [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Took 8.98 seconds to spawn the instance on the hypervisor. [ 979.929732] env[63021]: DEBUG nova.compute.manager [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 979.931033] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1205300-0af2-4dce-b4de-518081334561 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.941317] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fe550412-75b7-4078-8f53-e9fe3264ec95 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Releasing lock "refresh_cache-207c8197-6191-4837-8ead-8c24aa5b35bc" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.941317] env[63021]: DEBUG nova.objects.instance [None req-fe550412-75b7-4078-8f53-e9fe3264ec95 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lazy-loading 'migration_context' on Instance uuid 207c8197-6191-4837-8ead-8c24aa5b35bc {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 980.319543] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquiring lock "refresh_cache-df856146-61ba-4de7-a5fd-d75263927163" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.319846] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquired lock "refresh_cache-df856146-61ba-4de7-a5fd-d75263927163" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.320050] env[63021]: DEBUG nova.network.neutron [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: df856146-61ba-4de7-a5fd-d75263927163] Forcefully refreshing network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 980.448615] env[63021]: DEBUG nova.objects.base [None req-fe550412-75b7-4078-8f53-e9fe3264ec95 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Object Instance<207c8197-6191-4837-8ead-8c24aa5b35bc> lazy-loaded attributes: info_cache,migration_context {{(pid=63021) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 980.449579] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9808d886-c4cb-4ab3-850c-4761e9739dea {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.456421] env[63021]: INFO nova.compute.manager [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Took 20.90 seconds to build instance. [ 980.474297] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa0be55b-fdfb-40ca-b622-10173cb9557f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.481259] env[63021]: DEBUG oslo_vmware.api [None req-fe550412-75b7-4078-8f53-e9fe3264ec95 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 980.481259] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52030fa0-ab41-6fad-e435-9661c4f8d94f" [ 980.481259] env[63021]: _type = "Task" [ 980.481259] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.489961] env[63021]: DEBUG oslo_vmware.api [None req-fe550412-75b7-4078-8f53-e9fe3264ec95 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52030fa0-ab41-6fad-e435-9661c4f8d94f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.958979] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ac6e2952-ca0e-4144-af7f-245e7278ee57 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "67f736d9-82ec-448e-8b72-84fdd704aa8d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.414s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.992309] env[63021]: DEBUG oslo_vmware.api [None req-fe550412-75b7-4078-8f53-e9fe3264ec95 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52030fa0-ab41-6fad-e435-9661c4f8d94f, 'name': SearchDatastore_Task, 'duration_secs': 0.008256} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.992866] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fe550412-75b7-4078-8f53-e9fe3264ec95 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.993228] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fe550412-75b7-4078-8f53-e9fe3264ec95 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.348757] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "129d8c8a-b869-422d-8e7f-cc3a9400021a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.349137] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "129d8c8a-b869-422d-8e7f-cc3a9400021a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.693895] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a5e3291-1848-4d2c-a988-e17a9061cf94 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.701740] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b535700-61a9-43db-8fd8-b8ba61456780 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.742305] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9c24d3d-9e5e-43fa-9217-96d072ad732f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.750732] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdbcb76a-b988-4864-be6a-7b0073fbdac7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.768018] env[63021]: DEBUG nova.compute.provider_tree [None req-fe550412-75b7-4078-8f53-e9fe3264ec95 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 981.852635] env[63021]: DEBUG nova.compute.manager [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 982.020346] env[63021]: DEBUG nova.network.neutron [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: df856146-61ba-4de7-a5fd-d75263927163] Updating instance_info_cache with network_info: [{"id": "a43dcfd2-397e-44fe-bd1c-b6116c25a28a", "address": "fa:16:3e:c8:73:42", "network": {"id": "7258df3e-b1eb-4da6-ab52-5dadd2f87b51", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1084458843-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.144", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5661c44b534b4a7e9452e914fc80d66e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4f04accd-4a9f-41a1-b327-c96095e1ade4", "external-id": "nsx-vlan-transportzone-947", "segmentation_id": 947, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa43dcfd2-39", "ovs_interfaceid": "a43dcfd2-397e-44fe-bd1c-b6116c25a28a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.068884] env[63021]: DEBUG nova.compute.manager [req-e15932a6-caa2-4753-9101-7493e9d719d4 req-10004718-3c8f-455e-9dba-21bda5ebf82a service nova] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Received event network-changed-db333f88-b16f-439c-99fe-ca5884a09dc1 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 982.069116] env[63021]: DEBUG nova.compute.manager [req-e15932a6-caa2-4753-9101-7493e9d719d4 req-10004718-3c8f-455e-9dba-21bda5ebf82a service nova] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Refreshing instance network info cache due to event network-changed-db333f88-b16f-439c-99fe-ca5884a09dc1. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 982.069425] env[63021]: DEBUG oslo_concurrency.lockutils [req-e15932a6-caa2-4753-9101-7493e9d719d4 req-10004718-3c8f-455e-9dba-21bda5ebf82a service nova] Acquiring lock "refresh_cache-67f736d9-82ec-448e-8b72-84fdd704aa8d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.069509] env[63021]: DEBUG oslo_concurrency.lockutils [req-e15932a6-caa2-4753-9101-7493e9d719d4 req-10004718-3c8f-455e-9dba-21bda5ebf82a service nova] Acquired lock "refresh_cache-67f736d9-82ec-448e-8b72-84fdd704aa8d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.069700] env[63021]: DEBUG nova.network.neutron [req-e15932a6-caa2-4753-9101-7493e9d719d4 req-10004718-3c8f-455e-9dba-21bda5ebf82a service nova] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Refreshing network info cache for port db333f88-b16f-439c-99fe-ca5884a09dc1 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 982.272083] env[63021]: DEBUG nova.scheduler.client.report [None req-fe550412-75b7-4078-8f53-e9fe3264ec95 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 982.376569] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.524813] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Releasing lock "refresh_cache-df856146-61ba-4de7-a5fd-d75263927163" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.525086] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: df856146-61ba-4de7-a5fd-d75263927163] Updated the network info_cache for instance {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 982.525402] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 982.525452] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 982.525598] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 982.525748] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 982.525890] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 982.526041] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 982.526251] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63021) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 982.526422] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 982.802831] env[63021]: DEBUG nova.network.neutron [req-e15932a6-caa2-4753-9101-7493e9d719d4 req-10004718-3c8f-455e-9dba-21bda5ebf82a service nova] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Updated VIF entry in instance network info cache for port db333f88-b16f-439c-99fe-ca5884a09dc1. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 982.803292] env[63021]: DEBUG nova.network.neutron [req-e15932a6-caa2-4753-9101-7493e9d719d4 req-10004718-3c8f-455e-9dba-21bda5ebf82a service nova] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Updating instance_info_cache with network_info: [{"id": "db333f88-b16f-439c-99fe-ca5884a09dc1", "address": "fa:16:3e:40:f8:6d", "network": {"id": "3aaeed35-f452-4144-a259-8f440f8a8bf4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-229941363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.192", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b672b1d7d7ad454ebfa3e2fd2421c754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb333f88-b1", "ovs_interfaceid": "db333f88-b16f-439c-99fe-ca5884a09dc1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.030144] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.282106] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fe550412-75b7-4078-8f53-e9fe3264ec95 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.289s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.285837] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.909s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.286791] env[63021]: INFO nova.compute.claims [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 983.306299] env[63021]: DEBUG oslo_concurrency.lockutils [req-e15932a6-caa2-4753-9101-7493e9d719d4 req-10004718-3c8f-455e-9dba-21bda5ebf82a service nova] Releasing lock "refresh_cache-67f736d9-82ec-448e-8b72-84fdd704aa8d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.844095] env[63021]: INFO nova.scheduler.client.report [None req-fe550412-75b7-4078-8f53-e9fe3264ec95 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Deleted allocation for migration 3a372abc-b37d-4af2-9053-e4a6f57a8bbc [ 984.351703] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fe550412-75b7-4078-8f53-e9fe3264ec95 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "207c8197-6191-4837-8ead-8c24aa5b35bc" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.916s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.454826] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41df827b-bc95-41c1-acc2-68a8fd45263a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.462938] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-959735ae-125e-424d-870d-313146e68444 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.494510] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2caffecf-6c9f-4ad4-8d53-be8f42ad8d94 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.502303] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53de3bd6-b3ec-49ef-9535-1a126a60f7af {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.516286] env[63021]: DEBUG nova.compute.provider_tree [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 984.909747] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "207c8197-6191-4837-8ead-8c24aa5b35bc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.910172] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "207c8197-6191-4837-8ead-8c24aa5b35bc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.910226] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "207c8197-6191-4837-8ead-8c24aa5b35bc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.910423] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "207c8197-6191-4837-8ead-8c24aa5b35bc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.910600] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "207c8197-6191-4837-8ead-8c24aa5b35bc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.912816] env[63021]: INFO nova.compute.manager [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Terminating instance [ 984.914834] env[63021]: DEBUG nova.compute.manager [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 984.915048] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 984.915913] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83ff9cf5-4d43-4f1e-a0a9-4b5631053bf9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.924418] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 984.924751] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-958ea452-472e-4f07-9501-ffed6e860c2c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.931571] env[63021]: DEBUG oslo_vmware.api [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 984.931571] env[63021]: value = "task-1294119" [ 984.931571] env[63021]: _type = "Task" [ 984.931571] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.940701] env[63021]: DEBUG oslo_vmware.api [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294119, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.019636] env[63021]: DEBUG nova.scheduler.client.report [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 985.356554] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "9f822151-3307-4093-a972-8af128ea1892" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.356815] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "9f822151-3307-4093-a972-8af128ea1892" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.443131] env[63021]: DEBUG oslo_vmware.api [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294119, 'name': PowerOffVM_Task, 'duration_secs': 0.377017} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.443522] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 985.443749] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 985.444085] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3efb2497-4f0f-490d-b315-08be66dd9599 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.525759] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.241s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.526315] env[63021]: DEBUG nova.compute.manager [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 985.529544] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.500s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.529731] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.529881] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63021) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 985.530198] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 985.530386] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Deleting contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 985.530649] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Deleting the datastore file [datastore2] 207c8197-6191-4837-8ead-8c24aa5b35bc {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 985.531726] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8521dd6-2fd1-4c73-9fc4-018a4dbdb474 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.534468] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d9307c94-842e-48e2-987e-fe8006637b29 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.543592] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7e8475b-dc43-47f3-9db6-bc3674be3f3c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.547781] env[63021]: DEBUG oslo_vmware.api [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 985.547781] env[63021]: value = "task-1294121" [ 985.547781] env[63021]: _type = "Task" [ 985.547781] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.561543] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d77a8521-c030-4711-814e-39da4d768527 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.567579] env[63021]: DEBUG oslo_vmware.api [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294121, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.571975] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc2331d8-2e1b-426d-8443-51b0062356fe {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.603789] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180435MB free_disk=138GB free_vcpus=48 pci_devices=None {{(pid=63021) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 985.603986] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.604207] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.858875] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "interface-2d647028-306a-41d5-96f0-2e82a98ab56d-01e26b49-58a2-411d-b025-b1ca941026c4" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.859299] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "interface-2d647028-306a-41d5-96f0-2e82a98ab56d-01e26b49-58a2-411d-b025-b1ca941026c4" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.859814] env[63021]: DEBUG nova.objects.instance [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lazy-loading 'flavor' on Instance uuid 2d647028-306a-41d5-96f0-2e82a98ab56d {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 985.861836] env[63021]: DEBUG nova.compute.manager [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 986.032770] env[63021]: DEBUG nova.compute.utils [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 986.034236] env[63021]: DEBUG nova.compute.manager [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 986.034397] env[63021]: DEBUG nova.network.neutron [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 986.058477] env[63021]: DEBUG oslo_vmware.api [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294121, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.404094} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.058681] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 986.058737] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Deleted contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 986.058875] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 986.059145] env[63021]: INFO nova.compute.manager [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Took 1.14 seconds to destroy the instance on the hypervisor. [ 986.059320] env[63021]: DEBUG oslo.service.loopingcall [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 986.059509] env[63021]: DEBUG nova.compute.manager [-] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 986.059624] env[63021]: DEBUG nova.network.neutron [-] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 986.082503] env[63021]: DEBUG nova.policy [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ba16721213dd49a3b4e4990d7d319d1f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3c8959d2f1fe4706b7407c11e128b773', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 986.353549] env[63021]: DEBUG nova.compute.manager [req-eb4f102b-907c-40b6-947e-e5e619c9dbff req-263a9528-f8b4-4276-a8a6-2911589e46c8 service nova] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Received event network-vif-deleted-a7ecacae-44cb-4b30-892f-d0f5ef3ad9d5 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 986.353755] env[63021]: INFO nova.compute.manager [req-eb4f102b-907c-40b6-947e-e5e619c9dbff req-263a9528-f8b4-4276-a8a6-2911589e46c8 service nova] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Neutron deleted interface a7ecacae-44cb-4b30-892f-d0f5ef3ad9d5; detaching it from the instance and deleting it from the info cache [ 986.353967] env[63021]: DEBUG nova.network.neutron [req-eb4f102b-907c-40b6-947e-e5e619c9dbff req-263a9528-f8b4-4276-a8a6-2911589e46c8 service nova] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.392636] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.409530] env[63021]: DEBUG nova.network.neutron [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Successfully created port: 99e60085-e84c-4670-9b34-3ceb87f8db12 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 986.474721] env[63021]: DEBUG nova.objects.instance [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lazy-loading 'pci_requests' on Instance uuid 2d647028-306a-41d5-96f0-2e82a98ab56d {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 986.537884] env[63021]: DEBUG nova.compute.manager [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 986.645041] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance df856146-61ba-4de7-a5fd-d75263927163 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.645271] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance c328d2f7-3398-4f25-b11c-f464be7af8a1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.645426] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance ae103118-bddf-46fe-90b9-98b60952ebba actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.645565] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 418a394b-dbb2-47a2-96a9-a1caa4c4bb51 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.645769] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 2d647028-306a-41d5-96f0-2e82a98ab56d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.645838] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 45c9e2be-eb19-4b83-b280-c9eeaddfccba actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.645981] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 6155c098-e4a6-47e6-b343-4a77ca90eb2e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.646130] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 207c8197-6191-4837-8ead-8c24aa5b35bc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.646258] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance cee7a86d-91e2-4b57-b6cd-29056da6a414 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.646432] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 67f736d9-82ec-448e-8b72-84fdd704aa8d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.646680] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 129d8c8a-b869-422d-8e7f-cc3a9400021a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.832585] env[63021]: DEBUG nova.network.neutron [-] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.856985] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-72348587-8c11-42fc-9746-39385af4dce8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.866245] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76a82ff1-4686-4c60-8b13-588847a42535 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.896686] env[63021]: DEBUG nova.compute.manager [req-eb4f102b-907c-40b6-947e-e5e619c9dbff req-263a9528-f8b4-4276-a8a6-2911589e46c8 service nova] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Detach interface failed, port_id=a7ecacae-44cb-4b30-892f-d0f5ef3ad9d5, reason: Instance 207c8197-6191-4837-8ead-8c24aa5b35bc could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 986.977118] env[63021]: DEBUG nova.objects.base [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Object Instance<2d647028-306a-41d5-96f0-2e82a98ab56d> lazy-loaded attributes: flavor,pci_requests {{(pid=63021) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 986.977384] env[63021]: DEBUG nova.network.neutron [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 987.046729] env[63021]: DEBUG nova.policy [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6ea19453dbdb4251a93dde2576aeff64', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '58693d1e991249f28914a39761cc2627', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 987.152728] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 9f822151-3307-4093-a972-8af128ea1892 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 987.152974] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=63021) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 987.153321] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2688MB phys_disk=200GB used_disk=11GB total_vcpus=48 used_vcpus=11 pci_stats=[] {{(pid=63021) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 987.335219] env[63021]: INFO nova.compute.manager [-] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Took 1.28 seconds to deallocate network for instance. [ 987.339588] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-179eb326-9e48-4d83-97ad-1743d0a188f1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.351584] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08cbe9de-535c-4bdd-aa3b-8bb74c44242d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.385183] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cc632b1-24ee-4327-9c2b-198c820d5f29 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.393786] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9488b0c-1255-4ad1-b7b2-6eb6ee256680 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.409737] env[63021]: DEBUG nova.compute.provider_tree [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 987.428088] env[63021]: DEBUG oslo_vmware.rw_handles [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526e9ee5-d828-90b0-5e90-db179d7100a0/disk-0.vmdk. {{(pid=63021) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 987.428937] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bd8fc66-8b26-48e1-83d8-5f9d65f6cf79 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.435042] env[63021]: DEBUG oslo_vmware.rw_handles [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526e9ee5-d828-90b0-5e90-db179d7100a0/disk-0.vmdk is in state: ready. {{(pid=63021) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 987.435231] env[63021]: ERROR oslo_vmware.rw_handles [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526e9ee5-d828-90b0-5e90-db179d7100a0/disk-0.vmdk due to incomplete transfer. [ 987.435462] env[63021]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-9e7a8556-29ce-4b56-a32a-88ea18f974be {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.443051] env[63021]: DEBUG oslo_vmware.rw_handles [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526e9ee5-d828-90b0-5e90-db179d7100a0/disk-0.vmdk. {{(pid=63021) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 987.443250] env[63021]: DEBUG nova.virt.vmwareapi.images [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Uploaded image 9f7298ed-bbc9-476d-b57e-755a41852bfe to the Glance image server {{(pid=63021) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 987.445465] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Destroying the VM {{(pid=63021) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 987.445708] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-c5024413-98bd-4d46-a7a4-f055ba4c332f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.451641] env[63021]: DEBUG oslo_vmware.api [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 987.451641] env[63021]: value = "task-1294122" [ 987.451641] env[63021]: _type = "Task" [ 987.451641] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.458949] env[63021]: DEBUG oslo_vmware.api [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294122, 'name': Destroy_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.551297] env[63021]: DEBUG nova.compute.manager [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 987.572607] env[63021]: DEBUG nova.virt.hardware [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 987.572859] env[63021]: DEBUG nova.virt.hardware [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 987.573028] env[63021]: DEBUG nova.virt.hardware [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 987.573220] env[63021]: DEBUG nova.virt.hardware [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 987.573368] env[63021]: DEBUG nova.virt.hardware [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 987.573514] env[63021]: DEBUG nova.virt.hardware [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 987.573724] env[63021]: DEBUG nova.virt.hardware [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 987.573883] env[63021]: DEBUG nova.virt.hardware [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 987.574065] env[63021]: DEBUG nova.virt.hardware [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 987.574308] env[63021]: DEBUG nova.virt.hardware [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 987.574510] env[63021]: DEBUG nova.virt.hardware [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 987.575373] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b3f2ce8-5370-4ef1-a6b2-d89cba8dae4b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.583831] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83ee56d4-bfff-42b8-bffd-714c6a298674 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.847169] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.912911] env[63021]: DEBUG nova.scheduler.client.report [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 987.963402] env[63021]: DEBUG oslo_vmware.api [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294122, 'name': Destroy_Task, 'duration_secs': 0.325261} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.963703] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Destroyed the VM [ 987.963937] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Deleting Snapshot of the VM instance {{(pid=63021) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 987.964222] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-f37057f6-aa9c-462e-8035-3aa7871d7fe2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.970391] env[63021]: DEBUG oslo_vmware.api [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 987.970391] env[63021]: value = "task-1294123" [ 987.970391] env[63021]: _type = "Task" [ 987.970391] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.978808] env[63021]: DEBUG oslo_vmware.api [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294123, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.992257] env[63021]: DEBUG nova.compute.manager [req-bb4d02d3-e69b-4b60-aaca-3c9e3026e70e req-a258c23a-39ae-44f1-947f-12211b3a2d1a service nova] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Received event network-vif-plugged-99e60085-e84c-4670-9b34-3ceb87f8db12 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 987.992484] env[63021]: DEBUG oslo_concurrency.lockutils [req-bb4d02d3-e69b-4b60-aaca-3c9e3026e70e req-a258c23a-39ae-44f1-947f-12211b3a2d1a service nova] Acquiring lock "129d8c8a-b869-422d-8e7f-cc3a9400021a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.992694] env[63021]: DEBUG oslo_concurrency.lockutils [req-bb4d02d3-e69b-4b60-aaca-3c9e3026e70e req-a258c23a-39ae-44f1-947f-12211b3a2d1a service nova] Lock "129d8c8a-b869-422d-8e7f-cc3a9400021a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.992913] env[63021]: DEBUG oslo_concurrency.lockutils [req-bb4d02d3-e69b-4b60-aaca-3c9e3026e70e req-a258c23a-39ae-44f1-947f-12211b3a2d1a service nova] Lock "129d8c8a-b869-422d-8e7f-cc3a9400021a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.993045] env[63021]: DEBUG nova.compute.manager [req-bb4d02d3-e69b-4b60-aaca-3c9e3026e70e req-a258c23a-39ae-44f1-947f-12211b3a2d1a service nova] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] No waiting events found dispatching network-vif-plugged-99e60085-e84c-4670-9b34-3ceb87f8db12 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 987.993194] env[63021]: WARNING nova.compute.manager [req-bb4d02d3-e69b-4b60-aaca-3c9e3026e70e req-a258c23a-39ae-44f1-947f-12211b3a2d1a service nova] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Received unexpected event network-vif-plugged-99e60085-e84c-4670-9b34-3ceb87f8db12 for instance with vm_state building and task_state spawning. [ 988.059923] env[63021]: DEBUG nova.network.neutron [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Successfully updated port: 99e60085-e84c-4670-9b34-3ceb87f8db12 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 988.417081] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63021) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 988.417317] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.813s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.417607] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.025s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.419543] env[63021]: INFO nova.compute.claims [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 988.480044] env[63021]: DEBUG oslo_vmware.api [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294123, 'name': RemoveSnapshot_Task, 'duration_secs': 0.358957} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.480338] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Deleted Snapshot of the VM instance {{(pid=63021) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 988.480633] env[63021]: DEBUG nova.compute.manager [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 988.481447] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-680b4044-a093-4844-80a7-e8bcc51812b9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.487018] env[63021]: DEBUG nova.compute.manager [req-c92d3828-df40-498a-9106-99bea4ad7e6d req-bbfbb6d2-3c0b-495a-a23d-2b2f6846758e service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Received event network-vif-plugged-01e26b49-58a2-411d-b025-b1ca941026c4 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 988.487237] env[63021]: DEBUG oslo_concurrency.lockutils [req-c92d3828-df40-498a-9106-99bea4ad7e6d req-bbfbb6d2-3c0b-495a-a23d-2b2f6846758e service nova] Acquiring lock "2d647028-306a-41d5-96f0-2e82a98ab56d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.487538] env[63021]: DEBUG oslo_concurrency.lockutils [req-c92d3828-df40-498a-9106-99bea4ad7e6d req-bbfbb6d2-3c0b-495a-a23d-2b2f6846758e service nova] Lock "2d647028-306a-41d5-96f0-2e82a98ab56d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.487628] env[63021]: DEBUG oslo_concurrency.lockutils [req-c92d3828-df40-498a-9106-99bea4ad7e6d req-bbfbb6d2-3c0b-495a-a23d-2b2f6846758e service nova] Lock "2d647028-306a-41d5-96f0-2e82a98ab56d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.487761] env[63021]: DEBUG nova.compute.manager [req-c92d3828-df40-498a-9106-99bea4ad7e6d req-bbfbb6d2-3c0b-495a-a23d-2b2f6846758e service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] No waiting events found dispatching network-vif-plugged-01e26b49-58a2-411d-b025-b1ca941026c4 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 988.487926] env[63021]: WARNING nova.compute.manager [req-c92d3828-df40-498a-9106-99bea4ad7e6d req-bbfbb6d2-3c0b-495a-a23d-2b2f6846758e service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Received unexpected event network-vif-plugged-01e26b49-58a2-411d-b025-b1ca941026c4 for instance with vm_state active and task_state None. [ 988.562605] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "refresh_cache-129d8c8a-b869-422d-8e7f-cc3a9400021a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.562733] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquired lock "refresh_cache-129d8c8a-b869-422d-8e7f-cc3a9400021a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.562861] env[63021]: DEBUG nova.network.neutron [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 988.580624] env[63021]: DEBUG nova.network.neutron [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Successfully updated port: 01e26b49-58a2-411d-b025-b1ca941026c4 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 988.996649] env[63021]: INFO nova.compute.manager [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Shelve offloading [ 988.998416] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 988.998970] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1918d3ec-7834-4c36-ae30-56c58c584f55 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.007293] env[63021]: DEBUG oslo_vmware.api [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 989.007293] env[63021]: value = "task-1294124" [ 989.007293] env[63021]: _type = "Task" [ 989.007293] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.016647] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] VM already powered off {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 989.016850] env[63021]: DEBUG nova.compute.manager [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 989.017608] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73f973aa-5f1a-4980-b4b0-7940c9dd895e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.023265] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "refresh_cache-cee7a86d-91e2-4b57-b6cd-29056da6a414" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.023434] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquired lock "refresh_cache-cee7a86d-91e2-4b57-b6cd-29056da6a414" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.023604] env[63021]: DEBUG nova.network.neutron [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 989.085492] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "refresh_cache-2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.085808] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquired lock "refresh_cache-2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.085853] env[63021]: DEBUG nova.network.neutron [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 989.094933] env[63021]: DEBUG nova.network.neutron [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 989.221081] env[63021]: DEBUG nova.network.neutron [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Updating instance_info_cache with network_info: [{"id": "99e60085-e84c-4670-9b34-3ceb87f8db12", "address": "fa:16:3e:de:0d:5a", "network": {"id": "0c0dd1a7-6687-45c5-82b1-33f34bdbb848", "bridge": "br-int", "label": "tempest-ServersTestJSON-523240467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c8959d2f1fe4706b7407c11e128b773", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffc811e4-0e4e-4748-8c7e-b3f14ccbd42d", "external-id": "nsx-vlan-transportzone-404", "segmentation_id": 404, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99e60085-e8", "ovs_interfaceid": "99e60085-e84c-4670-9b34-3ceb87f8db12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.577919] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-662fce33-5214-42bb-813f-569ccd8596e6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.585420] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09d9996a-3ade-45c6-b7ba-d15666c0cf70 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.619812] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a93f2842-b2a0-44c2-a708-7e39fbf4487d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.622680] env[63021]: WARNING nova.network.neutron [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] 2c4f9111-2f39-4bd0-95bc-09b96d535ea3 already exists in list: networks containing: ['2c4f9111-2f39-4bd0-95bc-09b96d535ea3']. ignoring it [ 989.629093] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4cb0305-bb9d-492d-b711-b5c5a54ea6cf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.644807] env[63021]: DEBUG nova.compute.provider_tree [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 989.723775] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Releasing lock "refresh_cache-129d8c8a-b869-422d-8e7f-cc3a9400021a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.724087] env[63021]: DEBUG nova.compute.manager [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Instance network_info: |[{"id": "99e60085-e84c-4670-9b34-3ceb87f8db12", "address": "fa:16:3e:de:0d:5a", "network": {"id": "0c0dd1a7-6687-45c5-82b1-33f34bdbb848", "bridge": "br-int", "label": "tempest-ServersTestJSON-523240467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c8959d2f1fe4706b7407c11e128b773", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffc811e4-0e4e-4748-8c7e-b3f14ccbd42d", "external-id": "nsx-vlan-transportzone-404", "segmentation_id": 404, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99e60085-e8", "ovs_interfaceid": "99e60085-e84c-4670-9b34-3ceb87f8db12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 989.724491] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:de:0d:5a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ffc811e4-0e4e-4748-8c7e-b3f14ccbd42d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '99e60085-e84c-4670-9b34-3ceb87f8db12', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 989.732555] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Creating folder: Project (3c8959d2f1fe4706b7407c11e128b773). Parent ref: group-v277447. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 989.735228] env[63021]: DEBUG nova.network.neutron [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Updating instance_info_cache with network_info: [{"id": "7aec4a00-7fef-4c0f-a168-2cfc26dbe0be", "address": "fa:16:3e:f3:38:47", "network": {"id": "81917a55-4e41-4898-9f4e-4fced52b5a86", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1793837736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e238d5c1668540d2a9fffd3fd832d9b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7aec4a00-7f", "ovs_interfaceid": "7aec4a00-7fef-4c0f-a168-2cfc26dbe0be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.736357] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-94aeed96-3e79-4e9a-8437-112caae0976c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.749513] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Created folder: Project (3c8959d2f1fe4706b7407c11e128b773) in parent group-v277447. [ 989.749688] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Creating folder: Instances. Parent ref: group-v277574. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 989.749922] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b85b0416-c4e9-4cd9-b236-2962d7ee8c18 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.758412] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Created folder: Instances in parent group-v277574. [ 989.758676] env[63021]: DEBUG oslo.service.loopingcall [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 989.758889] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 989.759127] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4211e09f-c6ab-4b6c-99f1-58d2740dddab {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.779983] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 989.779983] env[63021]: value = "task-1294127" [ 989.779983] env[63021]: _type = "Task" [ 989.779983] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.787073] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294127, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.902156] env[63021]: DEBUG nova.network.neutron [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Updating instance_info_cache with network_info: [{"id": "eab7f4fe-4832-40e4-aed2-0e931d91a4d7", "address": "fa:16:3e:9c:24:88", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeab7f4fe-48", "ovs_interfaceid": "eab7f4fe-4832-40e4-aed2-0e931d91a4d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "01e26b49-58a2-411d-b025-b1ca941026c4", "address": "fa:16:3e:1d:da:30", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01e26b49-58", "ovs_interfaceid": "01e26b49-58a2-411d-b025-b1ca941026c4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.078789] env[63021]: DEBUG nova.compute.manager [req-73152d9f-f334-4ae4-9ac9-6a8abe8abfa3 req-8d87e2b6-ca50-45e4-a90b-af21177f2993 service nova] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Received event network-changed-99e60085-e84c-4670-9b34-3ceb87f8db12 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 990.079080] env[63021]: DEBUG nova.compute.manager [req-73152d9f-f334-4ae4-9ac9-6a8abe8abfa3 req-8d87e2b6-ca50-45e4-a90b-af21177f2993 service nova] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Refreshing instance network info cache due to event network-changed-99e60085-e84c-4670-9b34-3ceb87f8db12. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 990.079494] env[63021]: DEBUG oslo_concurrency.lockutils [req-73152d9f-f334-4ae4-9ac9-6a8abe8abfa3 req-8d87e2b6-ca50-45e4-a90b-af21177f2993 service nova] Acquiring lock "refresh_cache-129d8c8a-b869-422d-8e7f-cc3a9400021a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.079585] env[63021]: DEBUG oslo_concurrency.lockutils [req-73152d9f-f334-4ae4-9ac9-6a8abe8abfa3 req-8d87e2b6-ca50-45e4-a90b-af21177f2993 service nova] Acquired lock "refresh_cache-129d8c8a-b869-422d-8e7f-cc3a9400021a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.079729] env[63021]: DEBUG nova.network.neutron [req-73152d9f-f334-4ae4-9ac9-6a8abe8abfa3 req-8d87e2b6-ca50-45e4-a90b-af21177f2993 service nova] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Refreshing network info cache for port 99e60085-e84c-4670-9b34-3ceb87f8db12 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 990.149444] env[63021]: DEBUG nova.scheduler.client.report [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 990.239201] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Releasing lock "refresh_cache-cee7a86d-91e2-4b57-b6cd-29056da6a414" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.290081] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294127, 'name': CreateVM_Task, 'duration_secs': 0.321409} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.290245] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 990.290920] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.291103] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.291421] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 990.291678] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8177992c-6ddd-42ad-b52b-e4ffb5e0235f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.296644] env[63021]: DEBUG oslo_vmware.api [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 990.296644] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52780093-f645-414c-e4aa-6be7ad4248f6" [ 990.296644] env[63021]: _type = "Task" [ 990.296644] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.305740] env[63021]: DEBUG oslo_vmware.api [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52780093-f645-414c-e4aa-6be7ad4248f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.405066] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Releasing lock "refresh_cache-2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.405801] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.405966] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquired lock "2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.406910] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b70d85d-246b-4db8-a14e-f48bc5c84400 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.423693] env[63021]: DEBUG nova.virt.hardware [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 990.423931] env[63021]: DEBUG nova.virt.hardware [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 990.424101] env[63021]: DEBUG nova.virt.hardware [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 990.424288] env[63021]: DEBUG nova.virt.hardware [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 990.424434] env[63021]: DEBUG nova.virt.hardware [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 990.424585] env[63021]: DEBUG nova.virt.hardware [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 990.424786] env[63021]: DEBUG nova.virt.hardware [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 990.424940] env[63021]: DEBUG nova.virt.hardware [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 990.425116] env[63021]: DEBUG nova.virt.hardware [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 990.425279] env[63021]: DEBUG nova.virt.hardware [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 990.425448] env[63021]: DEBUG nova.virt.hardware [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 990.431709] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Reconfiguring VM to attach interface {{(pid=63021) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 990.431988] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-90b81774-b4a0-4eaa-8152-4f4b171d25cb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.449265] env[63021]: DEBUG oslo_vmware.api [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 990.449265] env[63021]: value = "task-1294128" [ 990.449265] env[63021]: _type = "Task" [ 990.449265] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.458495] env[63021]: DEBUG oslo_vmware.api [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294128, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.474795] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 990.475623] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63de684b-d2b7-41b6-935d-977429cfcd9f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.482351] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 990.482592] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-96abb5bf-d231-4a89-85be-c0e73dd186a2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.512609] env[63021]: DEBUG nova.compute.manager [req-b123eba0-2d33-4a63-ac17-b19745edfcde req-79bf6c97-ffa0-44f6-96fc-f39075ca6f99 service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Received event network-changed-01e26b49-58a2-411d-b025-b1ca941026c4 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 990.513021] env[63021]: DEBUG nova.compute.manager [req-b123eba0-2d33-4a63-ac17-b19745edfcde req-79bf6c97-ffa0-44f6-96fc-f39075ca6f99 service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Refreshing instance network info cache due to event network-changed-01e26b49-58a2-411d-b025-b1ca941026c4. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 990.513262] env[63021]: DEBUG oslo_concurrency.lockutils [req-b123eba0-2d33-4a63-ac17-b19745edfcde req-79bf6c97-ffa0-44f6-96fc-f39075ca6f99 service nova] Acquiring lock "refresh_cache-2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.513418] env[63021]: DEBUG oslo_concurrency.lockutils [req-b123eba0-2d33-4a63-ac17-b19745edfcde req-79bf6c97-ffa0-44f6-96fc-f39075ca6f99 service nova] Acquired lock "refresh_cache-2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.513585] env[63021]: DEBUG nova.network.neutron [req-b123eba0-2d33-4a63-ac17-b19745edfcde req-79bf6c97-ffa0-44f6-96fc-f39075ca6f99 service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Refreshing network info cache for port 01e26b49-58a2-411d-b025-b1ca941026c4 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 990.654109] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.236s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.654660] env[63021]: DEBUG nova.compute.manager [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 990.657750] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.811s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.657938] env[63021]: DEBUG nova.objects.instance [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lazy-loading 'resources' on Instance uuid 207c8197-6191-4837-8ead-8c24aa5b35bc {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 990.679635] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 990.679861] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 990.680159] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Deleting the datastore file [datastore1] cee7a86d-91e2-4b57-b6cd-29056da6a414 {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 990.680632] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e7035746-b925-4af1-84c4-0bfe8d2a2d25 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.688022] env[63021]: DEBUG oslo_vmware.api [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 990.688022] env[63021]: value = "task-1294130" [ 990.688022] env[63021]: _type = "Task" [ 990.688022] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.696386] env[63021]: DEBUG oslo_vmware.api [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294130, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.776880] env[63021]: DEBUG nova.network.neutron [req-73152d9f-f334-4ae4-9ac9-6a8abe8abfa3 req-8d87e2b6-ca50-45e4-a90b-af21177f2993 service nova] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Updated VIF entry in instance network info cache for port 99e60085-e84c-4670-9b34-3ceb87f8db12. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 990.777365] env[63021]: DEBUG nova.network.neutron [req-73152d9f-f334-4ae4-9ac9-6a8abe8abfa3 req-8d87e2b6-ca50-45e4-a90b-af21177f2993 service nova] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Updating instance_info_cache with network_info: [{"id": "99e60085-e84c-4670-9b34-3ceb87f8db12", "address": "fa:16:3e:de:0d:5a", "network": {"id": "0c0dd1a7-6687-45c5-82b1-33f34bdbb848", "bridge": "br-int", "label": "tempest-ServersTestJSON-523240467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c8959d2f1fe4706b7407c11e128b773", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffc811e4-0e4e-4748-8c7e-b3f14ccbd42d", "external-id": "nsx-vlan-transportzone-404", "segmentation_id": 404, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99e60085-e8", "ovs_interfaceid": "99e60085-e84c-4670-9b34-3ceb87f8db12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.807429] env[63021]: DEBUG oslo_vmware.api [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52780093-f645-414c-e4aa-6be7ad4248f6, 'name': SearchDatastore_Task, 'duration_secs': 0.009835} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.807740] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.807985] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 990.808233] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.808382] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.808561] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 990.808821] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d0a1753e-ff67-442a-81f2-2dd905f9ec5c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.825673] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 990.825858] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 990.826606] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-40263732-c08c-4530-a72b-2a6a7e4ac205 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.831673] env[63021]: DEBUG oslo_vmware.api [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 990.831673] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52309ae1-a4e0-5a09-d74a-d941c2e32d97" [ 990.831673] env[63021]: _type = "Task" [ 990.831673] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.839171] env[63021]: DEBUG oslo_vmware.api [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52309ae1-a4e0-5a09-d74a-d941c2e32d97, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.958467] env[63021]: DEBUG oslo_vmware.api [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294128, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.160695] env[63021]: DEBUG nova.compute.utils [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 991.165419] env[63021]: DEBUG nova.compute.manager [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 991.165592] env[63021]: DEBUG nova.network.neutron [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 991.198409] env[63021]: DEBUG oslo_vmware.api [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294130, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.229049} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.198674] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 991.198867] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 991.199318] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 991.208341] env[63021]: DEBUG nova.policy [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2642402689b04f89946802ac3b3e6287', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '805af233efa147b3982fb794cd9e6f73', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 991.220384] env[63021]: INFO nova.scheduler.client.report [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Deleted allocations for instance cee7a86d-91e2-4b57-b6cd-29056da6a414 [ 991.280017] env[63021]: DEBUG oslo_concurrency.lockutils [req-73152d9f-f334-4ae4-9ac9-6a8abe8abfa3 req-8d87e2b6-ca50-45e4-a90b-af21177f2993 service nova] Releasing lock "refresh_cache-129d8c8a-b869-422d-8e7f-cc3a9400021a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.281446] env[63021]: DEBUG nova.network.neutron [req-b123eba0-2d33-4a63-ac17-b19745edfcde req-79bf6c97-ffa0-44f6-96fc-f39075ca6f99 service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Updated VIF entry in instance network info cache for port 01e26b49-58a2-411d-b025-b1ca941026c4. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 991.281934] env[63021]: DEBUG nova.network.neutron [req-b123eba0-2d33-4a63-ac17-b19745edfcde req-79bf6c97-ffa0-44f6-96fc-f39075ca6f99 service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Updating instance_info_cache with network_info: [{"id": "eab7f4fe-4832-40e4-aed2-0e931d91a4d7", "address": "fa:16:3e:9c:24:88", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeab7f4fe-48", "ovs_interfaceid": "eab7f4fe-4832-40e4-aed2-0e931d91a4d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "01e26b49-58a2-411d-b025-b1ca941026c4", "address": "fa:16:3e:1d:da:30", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01e26b49-58", "ovs_interfaceid": "01e26b49-58a2-411d-b025-b1ca941026c4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.338890] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a2bfa3b-9725-499d-8af6-a917d26a8dc8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.347160] env[63021]: DEBUG oslo_vmware.api [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52309ae1-a4e0-5a09-d74a-d941c2e32d97, 'name': SearchDatastore_Task, 'duration_secs': 0.00884} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.349460] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ebbb72cf-7d65-4924-af54-09f1b63287ad {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.352508] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb3484d5-eeeb-4dd3-a1db-1ccf389b2f3e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.359599] env[63021]: DEBUG oslo_vmware.api [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 991.359599] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52bd45b9-740b-9036-eaf2-8f9555b56a0e" [ 991.359599] env[63021]: _type = "Task" [ 991.359599] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.389221] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ec96ec0-a979-41b8-aba0-e3b837ce3b05 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.398562] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4564dcf1-19ac-4f2a-b14c-da14b8d714a0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.402354] env[63021]: DEBUG oslo_vmware.api [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52bd45b9-740b-9036-eaf2-8f9555b56a0e, 'name': SearchDatastore_Task, 'duration_secs': 0.009428} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.402611] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.402858] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 129d8c8a-b869-422d-8e7f-cc3a9400021a/129d8c8a-b869-422d-8e7f-cc3a9400021a.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 991.403521] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5573f420-331a-4a9b-baaf-a390416cce12 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.420009] env[63021]: DEBUG nova.compute.provider_tree [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 991.426450] env[63021]: DEBUG oslo_vmware.api [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 991.426450] env[63021]: value = "task-1294131" [ 991.426450] env[63021]: _type = "Task" [ 991.426450] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.435192] env[63021]: DEBUG oslo_vmware.api [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294131, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.459575] env[63021]: DEBUG oslo_vmware.api [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294128, 'name': ReconfigVM_Task, 'duration_secs': 0.695593} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.460131] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Releasing lock "2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.460328] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Reconfigured VM to attach interface {{(pid=63021) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 991.467208] env[63021]: DEBUG nova.network.neutron [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Successfully created port: e5cc5b77-3f1d-444f-bb24-2d7f98b67a8b {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 991.669064] env[63021]: DEBUG nova.compute.manager [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 991.725416] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.784804] env[63021]: DEBUG oslo_concurrency.lockutils [req-b123eba0-2d33-4a63-ac17-b19745edfcde req-79bf6c97-ffa0-44f6-96fc-f39075ca6f99 service nova] Releasing lock "refresh_cache-2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.923203] env[63021]: DEBUG nova.scheduler.client.report [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 991.937134] env[63021]: DEBUG oslo_vmware.api [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294131, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.476} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.938050] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 129d8c8a-b869-422d-8e7f-cc3a9400021a/129d8c8a-b869-422d-8e7f-cc3a9400021a.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 991.938279] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 991.938613] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a35f6401-7de3-4594-bae0-9fa7320f88dd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.945138] env[63021]: DEBUG oslo_vmware.api [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 991.945138] env[63021]: value = "task-1294132" [ 991.945138] env[63021]: _type = "Task" [ 991.945138] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.953364] env[63021]: DEBUG oslo_vmware.api [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294132, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.965138] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3d5ed7cc-9d61-4982-ad59-d1e91ccb0612 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "interface-2d647028-306a-41d5-96f0-2e82a98ab56d-01e26b49-58a2-411d-b025-b1ca941026c4" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.106s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.109738] env[63021]: DEBUG nova.compute.manager [req-70161a65-3685-47a3-af32-697292dde960 req-ae197729-8da9-4029-ac6e-8c4811031a71 service nova] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Received event network-vif-unplugged-7aec4a00-7fef-4c0f-a168-2cfc26dbe0be {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 992.110027] env[63021]: DEBUG oslo_concurrency.lockutils [req-70161a65-3685-47a3-af32-697292dde960 req-ae197729-8da9-4029-ac6e-8c4811031a71 service nova] Acquiring lock "cee7a86d-91e2-4b57-b6cd-29056da6a414-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.110243] env[63021]: DEBUG oslo_concurrency.lockutils [req-70161a65-3685-47a3-af32-697292dde960 req-ae197729-8da9-4029-ac6e-8c4811031a71 service nova] Lock "cee7a86d-91e2-4b57-b6cd-29056da6a414-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.110413] env[63021]: DEBUG oslo_concurrency.lockutils [req-70161a65-3685-47a3-af32-697292dde960 req-ae197729-8da9-4029-ac6e-8c4811031a71 service nova] Lock "cee7a86d-91e2-4b57-b6cd-29056da6a414-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.110636] env[63021]: DEBUG nova.compute.manager [req-70161a65-3685-47a3-af32-697292dde960 req-ae197729-8da9-4029-ac6e-8c4811031a71 service nova] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] No waiting events found dispatching network-vif-unplugged-7aec4a00-7fef-4c0f-a168-2cfc26dbe0be {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 992.110827] env[63021]: WARNING nova.compute.manager [req-70161a65-3685-47a3-af32-697292dde960 req-ae197729-8da9-4029-ac6e-8c4811031a71 service nova] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Received unexpected event network-vif-unplugged-7aec4a00-7fef-4c0f-a168-2cfc26dbe0be for instance with vm_state shelved_offloaded and task_state None. [ 992.110992] env[63021]: DEBUG nova.compute.manager [req-70161a65-3685-47a3-af32-697292dde960 req-ae197729-8da9-4029-ac6e-8c4811031a71 service nova] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Received event network-changed-7aec4a00-7fef-4c0f-a168-2cfc26dbe0be {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 992.111165] env[63021]: DEBUG nova.compute.manager [req-70161a65-3685-47a3-af32-697292dde960 req-ae197729-8da9-4029-ac6e-8c4811031a71 service nova] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Refreshing instance network info cache due to event network-changed-7aec4a00-7fef-4c0f-a168-2cfc26dbe0be. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 992.111347] env[63021]: DEBUG oslo_concurrency.lockutils [req-70161a65-3685-47a3-af32-697292dde960 req-ae197729-8da9-4029-ac6e-8c4811031a71 service nova] Acquiring lock "refresh_cache-cee7a86d-91e2-4b57-b6cd-29056da6a414" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.111486] env[63021]: DEBUG oslo_concurrency.lockutils [req-70161a65-3685-47a3-af32-697292dde960 req-ae197729-8da9-4029-ac6e-8c4811031a71 service nova] Acquired lock "refresh_cache-cee7a86d-91e2-4b57-b6cd-29056da6a414" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.111644] env[63021]: DEBUG nova.network.neutron [req-70161a65-3685-47a3-af32-697292dde960 req-ae197729-8da9-4029-ac6e-8c4811031a71 service nova] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Refreshing network info cache for port 7aec4a00-7fef-4c0f-a168-2cfc26dbe0be {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 992.428203] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.770s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.430893] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.705s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.430893] env[63021]: DEBUG nova.objects.instance [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lazy-loading 'resources' on Instance uuid cee7a86d-91e2-4b57-b6cd-29056da6a414 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 992.458308] env[63021]: DEBUG oslo_vmware.api [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294132, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.12685} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.458308] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 992.458308] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cce5e7bf-8812-442c-926e-83cc704fca95 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.482231] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 129d8c8a-b869-422d-8e7f-cc3a9400021a/129d8c8a-b869-422d-8e7f-cc3a9400021a.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 992.483464] env[63021]: INFO nova.scheduler.client.report [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Deleted allocations for instance 207c8197-6191-4837-8ead-8c24aa5b35bc [ 992.485447] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2f9b5a18-749e-4a84-ac7b-c1e80b21603b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.508853] env[63021]: DEBUG oslo_vmware.api [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 992.508853] env[63021]: value = "task-1294133" [ 992.508853] env[63021]: _type = "Task" [ 992.508853] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.516701] env[63021]: DEBUG oslo_vmware.api [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294133, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.679541] env[63021]: DEBUG nova.compute.manager [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 992.700079] env[63021]: DEBUG nova.virt.hardware [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 992.700439] env[63021]: DEBUG nova.virt.hardware [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 992.700620] env[63021]: DEBUG nova.virt.hardware [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 992.700749] env[63021]: DEBUG nova.virt.hardware [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 992.700901] env[63021]: DEBUG nova.virt.hardware [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 992.701060] env[63021]: DEBUG nova.virt.hardware [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 992.701273] env[63021]: DEBUG nova.virt.hardware [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 992.701434] env[63021]: DEBUG nova.virt.hardware [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 992.701604] env[63021]: DEBUG nova.virt.hardware [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 992.701799] env[63021]: DEBUG nova.virt.hardware [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 992.701988] env[63021]: DEBUG nova.virt.hardware [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 992.702843] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a145c22-a35a-4334-9d31-6afd53cb9e0a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.710802] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be9cb1b3-d583-4dbc-9734-39385279cb08 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.914016] env[63021]: DEBUG nova.network.neutron [req-70161a65-3685-47a3-af32-697292dde960 req-ae197729-8da9-4029-ac6e-8c4811031a71 service nova] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Updated VIF entry in instance network info cache for port 7aec4a00-7fef-4c0f-a168-2cfc26dbe0be. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 992.914516] env[63021]: DEBUG nova.network.neutron [req-70161a65-3685-47a3-af32-697292dde960 req-ae197729-8da9-4029-ac6e-8c4811031a71 service nova] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Updating instance_info_cache with network_info: [{"id": "7aec4a00-7fef-4c0f-a168-2cfc26dbe0be", "address": "fa:16:3e:f3:38:47", "network": {"id": "81917a55-4e41-4898-9f4e-4fced52b5a86", "bridge": null, "label": "tempest-DeleteServersTestJSON-1793837736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e238d5c1668540d2a9fffd3fd832d9b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap7aec4a00-7f", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.933865] env[63021]: DEBUG nova.objects.instance [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lazy-loading 'numa_topology' on Instance uuid cee7a86d-91e2-4b57-b6cd-29056da6a414 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 992.944957] env[63021]: DEBUG nova.compute.manager [req-54f1b613-53ca-4365-95d5-9f3165dc3273 req-957e0fc9-f1d7-49c6-b8ab-7fdbd2aa42bc service nova] [instance: 9f822151-3307-4093-a972-8af128ea1892] Received event network-vif-plugged-e5cc5b77-3f1d-444f-bb24-2d7f98b67a8b {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 992.945670] env[63021]: DEBUG oslo_concurrency.lockutils [req-54f1b613-53ca-4365-95d5-9f3165dc3273 req-957e0fc9-f1d7-49c6-b8ab-7fdbd2aa42bc service nova] Acquiring lock "9f822151-3307-4093-a972-8af128ea1892-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.945964] env[63021]: DEBUG oslo_concurrency.lockutils [req-54f1b613-53ca-4365-95d5-9f3165dc3273 req-957e0fc9-f1d7-49c6-b8ab-7fdbd2aa42bc service nova] Lock "9f822151-3307-4093-a972-8af128ea1892-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.946177] env[63021]: DEBUG oslo_concurrency.lockutils [req-54f1b613-53ca-4365-95d5-9f3165dc3273 req-957e0fc9-f1d7-49c6-b8ab-7fdbd2aa42bc service nova] Lock "9f822151-3307-4093-a972-8af128ea1892-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.946415] env[63021]: DEBUG nova.compute.manager [req-54f1b613-53ca-4365-95d5-9f3165dc3273 req-957e0fc9-f1d7-49c6-b8ab-7fdbd2aa42bc service nova] [instance: 9f822151-3307-4093-a972-8af128ea1892] No waiting events found dispatching network-vif-plugged-e5cc5b77-3f1d-444f-bb24-2d7f98b67a8b {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 992.946602] env[63021]: WARNING nova.compute.manager [req-54f1b613-53ca-4365-95d5-9f3165dc3273 req-957e0fc9-f1d7-49c6-b8ab-7fdbd2aa42bc service nova] [instance: 9f822151-3307-4093-a972-8af128ea1892] Received unexpected event network-vif-plugged-e5cc5b77-3f1d-444f-bb24-2d7f98b67a8b for instance with vm_state building and task_state spawning. [ 993.001055] env[63021]: DEBUG nova.network.neutron [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Successfully updated port: e5cc5b77-3f1d-444f-bb24-2d7f98b67a8b {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 993.006030] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3ac59fd3-7b07-4401-94d2-5287bc09ad59 tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "207c8197-6191-4837-8ead-8c24aa5b35bc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.096s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.019144] env[63021]: DEBUG oslo_vmware.api [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294133, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.102047] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2818b789-eae5-4b0e-b064-a0fbb9dbbc9f tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "cee7a86d-91e2-4b57-b6cd-29056da6a414" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.418425] env[63021]: DEBUG oslo_concurrency.lockutils [req-70161a65-3685-47a3-af32-697292dde960 req-ae197729-8da9-4029-ac6e-8c4811031a71 service nova] Releasing lock "refresh_cache-cee7a86d-91e2-4b57-b6cd-29056da6a414" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.436257] env[63021]: DEBUG nova.objects.base [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=63021) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 993.504779] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "refresh_cache-9f822151-3307-4093-a972-8af128ea1892" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.505475] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquired lock "refresh_cache-9f822151-3307-4093-a972-8af128ea1892" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.507343] env[63021]: DEBUG nova.network.neutron [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 993.522266] env[63021]: DEBUG oslo_vmware.api [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294133, 'name': ReconfigVM_Task, 'duration_secs': 0.818407} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.524710] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 129d8c8a-b869-422d-8e7f-cc3a9400021a/129d8c8a-b869-422d-8e7f-cc3a9400021a.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 993.525492] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-950e0967-277c-4c40-9b94-e14128e6802f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.531663] env[63021]: DEBUG oslo_vmware.api [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 993.531663] env[63021]: value = "task-1294134" [ 993.531663] env[63021]: _type = "Task" [ 993.531663] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.540940] env[63021]: DEBUG oslo_vmware.api [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294134, 'name': Rename_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.560764] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b8fcad9f-92f2-40f7-bc7d-c79c2605d1cb tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "interface-2d647028-306a-41d5-96f0-2e82a98ab56d-01e26b49-58a2-411d-b025-b1ca941026c4" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.561056] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b8fcad9f-92f2-40f7-bc7d-c79c2605d1cb tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "interface-2d647028-306a-41d5-96f0-2e82a98ab56d-01e26b49-58a2-411d-b025-b1ca941026c4" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.569785] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f109f88b-c557-446c-a20f-dbbf043976f3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.578374] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-511aa559-a236-46d6-b0d5-729462725f32 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.609358] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fc91380-848e-4328-9fc7-f5a63973950a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.616806] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57130bf0-e4e2-420e-b6ac-85911380c678 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.631029] env[63021]: DEBUG nova.compute.provider_tree [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 994.038275] env[63021]: DEBUG nova.network.neutron [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 994.043408] env[63021]: DEBUG oslo_vmware.api [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294134, 'name': Rename_Task, 'duration_secs': 0.138634} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.043723] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 994.043986] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b9934332-0e4c-402f-8fde-c4218f2b4588 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.049689] env[63021]: DEBUG oslo_vmware.api [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 994.049689] env[63021]: value = "task-1294135" [ 994.049689] env[63021]: _type = "Task" [ 994.049689] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.057698] env[63021]: DEBUG oslo_vmware.api [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294135, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.063261] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b8fcad9f-92f2-40f7-bc7d-c79c2605d1cb tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.063431] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b8fcad9f-92f2-40f7-bc7d-c79c2605d1cb tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquired lock "2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.064163] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab684bd5-553c-4ce3-b9da-25b8ba936132 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.082041] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e989757-e8a8-4095-860a-f8c8efb8e0f0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.109789] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b8fcad9f-92f2-40f7-bc7d-c79c2605d1cb tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Reconfiguring VM to detach interface {{(pid=63021) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 994.110110] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fe2c9516-01cd-4956-82ad-7092b696c1ef {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.130587] env[63021]: DEBUG oslo_vmware.api [None req-b8fcad9f-92f2-40f7-bc7d-c79c2605d1cb tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 994.130587] env[63021]: value = "task-1294136" [ 994.130587] env[63021]: _type = "Task" [ 994.130587] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.134485] env[63021]: DEBUG nova.scheduler.client.report [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 994.144929] env[63021]: DEBUG oslo_vmware.api [None req-b8fcad9f-92f2-40f7-bc7d-c79c2605d1cb tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294136, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.198503] env[63021]: DEBUG nova.network.neutron [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Updating instance_info_cache with network_info: [{"id": "e5cc5b77-3f1d-444f-bb24-2d7f98b67a8b", "address": "fa:16:3e:cd:30:f5", "network": {"id": "708d0dc3-c651-4001-9638-1d3538ff76c9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-983471056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "805af233efa147b3982fb794cd9e6f73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5cc5b77-3f", "ovs_interfaceid": "e5cc5b77-3f1d-444f-bb24-2d7f98b67a8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.559846] env[63021]: DEBUG oslo_vmware.api [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294135, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.641337] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.211s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.643794] env[63021]: DEBUG oslo_vmware.api [None req-b8fcad9f-92f2-40f7-bc7d-c79c2605d1cb tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294136, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.701654] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Releasing lock "refresh_cache-9f822151-3307-4093-a972-8af128ea1892" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.701654] env[63021]: DEBUG nova.compute.manager [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Instance network_info: |[{"id": "e5cc5b77-3f1d-444f-bb24-2d7f98b67a8b", "address": "fa:16:3e:cd:30:f5", "network": {"id": "708d0dc3-c651-4001-9638-1d3538ff76c9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-983471056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "805af233efa147b3982fb794cd9e6f73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5cc5b77-3f", "ovs_interfaceid": "e5cc5b77-3f1d-444f-bb24-2d7f98b67a8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 994.702196] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cd:30:f5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b7a73c01-1bb9-4612-a1a7-16d71b732e81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e5cc5b77-3f1d-444f-bb24-2d7f98b67a8b', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 994.710606] env[63021]: DEBUG oslo.service.loopingcall [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 994.710828] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f822151-3307-4093-a972-8af128ea1892] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 994.711099] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c4399044-b6b7-4a97-8e23-27006d6fa70a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.730908] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 994.730908] env[63021]: value = "task-1294137" [ 994.730908] env[63021]: _type = "Task" [ 994.730908] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.738359] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294137, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.970965] env[63021]: DEBUG nova.compute.manager [req-a7b03713-d676-47c4-a2d6-31037da00c2d req-b04f2414-5fa7-4c93-850b-301825403596 service nova] [instance: 9f822151-3307-4093-a972-8af128ea1892] Received event network-changed-e5cc5b77-3f1d-444f-bb24-2d7f98b67a8b {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 994.971063] env[63021]: DEBUG nova.compute.manager [req-a7b03713-d676-47c4-a2d6-31037da00c2d req-b04f2414-5fa7-4c93-850b-301825403596 service nova] [instance: 9f822151-3307-4093-a972-8af128ea1892] Refreshing instance network info cache due to event network-changed-e5cc5b77-3f1d-444f-bb24-2d7f98b67a8b. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 994.971273] env[63021]: DEBUG oslo_concurrency.lockutils [req-a7b03713-d676-47c4-a2d6-31037da00c2d req-b04f2414-5fa7-4c93-850b-301825403596 service nova] Acquiring lock "refresh_cache-9f822151-3307-4093-a972-8af128ea1892" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.971420] env[63021]: DEBUG oslo_concurrency.lockutils [req-a7b03713-d676-47c4-a2d6-31037da00c2d req-b04f2414-5fa7-4c93-850b-301825403596 service nova] Acquired lock "refresh_cache-9f822151-3307-4093-a972-8af128ea1892" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.971585] env[63021]: DEBUG nova.network.neutron [req-a7b03713-d676-47c4-a2d6-31037da00c2d req-b04f2414-5fa7-4c93-850b-301825403596 service nova] [instance: 9f822151-3307-4093-a972-8af128ea1892] Refreshing network info cache for port e5cc5b77-3f1d-444f-bb24-2d7f98b67a8b {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 995.059586] env[63021]: DEBUG oslo_vmware.api [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294135, 'name': PowerOnVM_Task, 'duration_secs': 0.527901} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.059804] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 995.060046] env[63021]: INFO nova.compute.manager [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Took 7.51 seconds to spawn the instance on the hypervisor. [ 995.060239] env[63021]: DEBUG nova.compute.manager [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 995.061045] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb5739cf-e935-41b6-83a8-4544d5851d80 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.140278] env[63021]: DEBUG oslo_vmware.api [None req-b8fcad9f-92f2-40f7-bc7d-c79c2605d1cb tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294136, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.150506] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1b771ea8-6bd0-4cb1-8d3a-8f866c44d343 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "cee7a86d-91e2-4b57-b6cd-29056da6a414" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 20.641s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.151438] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2818b789-eae5-4b0e-b064-a0fbb9dbbc9f tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "cee7a86d-91e2-4b57-b6cd-29056da6a414" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 2.049s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.151519] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2818b789-eae5-4b0e-b064-a0fbb9dbbc9f tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "cee7a86d-91e2-4b57-b6cd-29056da6a414-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.151924] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2818b789-eae5-4b0e-b064-a0fbb9dbbc9f tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "cee7a86d-91e2-4b57-b6cd-29056da6a414-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.152128] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2818b789-eae5-4b0e-b064-a0fbb9dbbc9f tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "cee7a86d-91e2-4b57-b6cd-29056da6a414-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.153756] env[63021]: INFO nova.compute.manager [None req-2818b789-eae5-4b0e-b064-a0fbb9dbbc9f tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Terminating instance [ 995.155541] env[63021]: DEBUG nova.compute.manager [None req-2818b789-eae5-4b0e-b064-a0fbb9dbbc9f tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 995.155784] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2818b789-eae5-4b0e-b064-a0fbb9dbbc9f tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 995.156091] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bb582e96-7f4a-48df-b5f1-92a01d29d9f4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.165126] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1efe9568-a229-4fd4-a296-3e2d226bbc1a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.194569] env[63021]: WARNING nova.virt.vmwareapi.vmops [None req-2818b789-eae5-4b0e-b064-a0fbb9dbbc9f tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance cee7a86d-91e2-4b57-b6cd-29056da6a414 could not be found. [ 995.194802] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2818b789-eae5-4b0e-b064-a0fbb9dbbc9f tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 995.194983] env[63021]: INFO nova.compute.manager [None req-2818b789-eae5-4b0e-b064-a0fbb9dbbc9f tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Took 0.04 seconds to destroy the instance on the hypervisor. [ 995.195248] env[63021]: DEBUG oslo.service.loopingcall [None req-2818b789-eae5-4b0e-b064-a0fbb9dbbc9f tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 995.195709] env[63021]: DEBUG nova.compute.manager [-] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 995.195709] env[63021]: DEBUG nova.network.neutron [-] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 995.240877] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294137, 'name': CreateVM_Task, 'duration_secs': 0.330757} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.240993] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f822151-3307-4093-a972-8af128ea1892] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 995.241652] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.241822] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.242143] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 995.242417] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef1eeae2-3220-40da-b70c-2bfea1ae20b8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.246871] env[63021]: DEBUG oslo_vmware.api [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 995.246871] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52628b28-aaf0-7e3c-4a1f-29095cb5ca74" [ 995.246871] env[63021]: _type = "Task" [ 995.246871] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.255949] env[63021]: DEBUG oslo_vmware.api [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52628b28-aaf0-7e3c-4a1f-29095cb5ca74, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.577523] env[63021]: INFO nova.compute.manager [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Took 13.22 seconds to build instance. [ 995.641631] env[63021]: DEBUG oslo_vmware.api [None req-b8fcad9f-92f2-40f7-bc7d-c79c2605d1cb tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294136, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.673441] env[63021]: DEBUG nova.network.neutron [req-a7b03713-d676-47c4-a2d6-31037da00c2d req-b04f2414-5fa7-4c93-850b-301825403596 service nova] [instance: 9f822151-3307-4093-a972-8af128ea1892] Updated VIF entry in instance network info cache for port e5cc5b77-3f1d-444f-bb24-2d7f98b67a8b. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 995.673856] env[63021]: DEBUG nova.network.neutron [req-a7b03713-d676-47c4-a2d6-31037da00c2d req-b04f2414-5fa7-4c93-850b-301825403596 service nova] [instance: 9f822151-3307-4093-a972-8af128ea1892] Updating instance_info_cache with network_info: [{"id": "e5cc5b77-3f1d-444f-bb24-2d7f98b67a8b", "address": "fa:16:3e:cd:30:f5", "network": {"id": "708d0dc3-c651-4001-9638-1d3538ff76c9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-983471056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "805af233efa147b3982fb794cd9e6f73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5cc5b77-3f", "ovs_interfaceid": "e5cc5b77-3f1d-444f-bb24-2d7f98b67a8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.757823] env[63021]: DEBUG oslo_vmware.api [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52628b28-aaf0-7e3c-4a1f-29095cb5ca74, 'name': SearchDatastore_Task, 'duration_secs': 0.008906} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.758143] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.758375] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 995.758618] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.758764] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.758938] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 995.759212] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7e1a92e1-a3fc-4aa8-940c-d00c45602e49 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.767229] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 995.767397] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 995.768077] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b64f622d-497b-400a-99be-5350eae42420 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.772531] env[63021]: DEBUG oslo_vmware.api [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 995.772531] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52204ee4-1a0a-e247-7f8d-e84a0aadea01" [ 995.772531] env[63021]: _type = "Task" [ 995.772531] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.779641] env[63021]: DEBUG oslo_vmware.api [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52204ee4-1a0a-e247-7f8d-e84a0aadea01, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.916662] env[63021]: DEBUG nova.network.neutron [-] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.993770] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "67aed8eb-d08c-47cc-8558-5c5a0f058d6e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.994015] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "67aed8eb-d08c-47cc-8558-5c5a0f058d6e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.080025] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d600586c-e0a7-4bd8-8bb8-1ad57d5872c1 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "129d8c8a-b869-422d-8e7f-cc3a9400021a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.731s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.143673] env[63021]: DEBUG oslo_vmware.api [None req-b8fcad9f-92f2-40f7-bc7d-c79c2605d1cb tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294136, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.176539] env[63021]: DEBUG oslo_concurrency.lockutils [req-a7b03713-d676-47c4-a2d6-31037da00c2d req-b04f2414-5fa7-4c93-850b-301825403596 service nova] Releasing lock "refresh_cache-9f822151-3307-4093-a972-8af128ea1892" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.282996] env[63021]: DEBUG oslo_vmware.api [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52204ee4-1a0a-e247-7f8d-e84a0aadea01, 'name': SearchDatastore_Task, 'duration_secs': 0.007826} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.283791] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a14f26c-82b0-4aac-8f9b-951e18efdbf8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.288925] env[63021]: DEBUG oslo_vmware.api [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 996.288925] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5205cb3a-85e4-f397-2bbc-89bfc7a520b6" [ 996.288925] env[63021]: _type = "Task" [ 996.288925] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.296473] env[63021]: DEBUG oslo_vmware.api [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5205cb3a-85e4-f397-2bbc-89bfc7a520b6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.419865] env[63021]: INFO nova.compute.manager [-] [instance: cee7a86d-91e2-4b57-b6cd-29056da6a414] Took 1.22 seconds to deallocate network for instance. [ 996.496298] env[63021]: DEBUG nova.compute.manager [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 996.642967] env[63021]: DEBUG oslo_vmware.api [None req-b8fcad9f-92f2-40f7-bc7d-c79c2605d1cb tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294136, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.801354] env[63021]: DEBUG oslo_vmware.api [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5205cb3a-85e4-f397-2bbc-89bfc7a520b6, 'name': SearchDatastore_Task, 'duration_secs': 0.008862} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.801750] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.802042] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 9f822151-3307-4093-a972-8af128ea1892/9f822151-3307-4093-a972-8af128ea1892.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 996.802310] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-90bbcd20-c6b7-46fc-8996-403ba75ed498 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.809357] env[63021]: DEBUG oslo_vmware.api [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 996.809357] env[63021]: value = "task-1294138" [ 996.809357] env[63021]: _type = "Task" [ 996.809357] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.817437] env[63021]: DEBUG oslo_vmware.api [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294138, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.016663] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.016663] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.019180] env[63021]: INFO nova.compute.claims [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 997.144083] env[63021]: DEBUG oslo_vmware.api [None req-b8fcad9f-92f2-40f7-bc7d-c79c2605d1cb tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294136, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.319242] env[63021]: DEBUG oslo_vmware.api [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294138, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.463853} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.319512] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 9f822151-3307-4093-a972-8af128ea1892/9f822151-3307-4093-a972-8af128ea1892.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 997.319759] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 997.320351] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b203028a-37ef-4c6d-877d-d5fbe81f3921 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.326091] env[63021]: DEBUG oslo_vmware.api [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 997.326091] env[63021]: value = "task-1294139" [ 997.326091] env[63021]: _type = "Task" [ 997.326091] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.333487] env[63021]: DEBUG oslo_vmware.api [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294139, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.445971] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2818b789-eae5-4b0e-b064-a0fbb9dbbc9f tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "cee7a86d-91e2-4b57-b6cd-29056da6a414" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.295s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.555675] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "023c0891-2b08-48bb-9754-408ce7d04c9c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.555941] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "023c0891-2b08-48bb-9754-408ce7d04c9c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.644827] env[63021]: DEBUG oslo_vmware.api [None req-b8fcad9f-92f2-40f7-bc7d-c79c2605d1cb tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294136, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.835964] env[63021]: DEBUG oslo_vmware.api [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294139, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.109641} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.836367] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 997.838040] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cf22221-c71d-4739-b4ca-f7d78b8a6eb1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.860936] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] 9f822151-3307-4093-a972-8af128ea1892/9f822151-3307-4093-a972-8af128ea1892.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 997.861241] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-245f77d0-95af-41d7-946c-40d6ad235844 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.879815] env[63021]: DEBUG oslo_vmware.api [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 997.879815] env[63021]: value = "task-1294140" [ 997.879815] env[63021]: _type = "Task" [ 997.879815] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.887368] env[63021]: DEBUG oslo_vmware.api [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294140, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.987238] env[63021]: DEBUG oslo_concurrency.lockutils [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "07510760-13ff-4797-86ca-86b384aacdfa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.987510] env[63021]: DEBUG oslo_concurrency.lockutils [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "07510760-13ff-4797-86ca-86b384aacdfa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.059161] env[63021]: DEBUG nova.compute.manager [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 998.146955] env[63021]: DEBUG oslo_vmware.api [None req-b8fcad9f-92f2-40f7-bc7d-c79c2605d1cb tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294136, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.210278] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ddbda57-9ced-4c61-90b7-7fceac754b42 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.217896] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea4e5fa3-6285-4242-bf71-cdb5774c7f61 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.248656] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e1df690-615b-4d06-9199-e3ee1b93856b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.255977] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-584cc1de-485d-4eba-81fd-beae35083384 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.268947] env[63021]: DEBUG nova.compute.provider_tree [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 998.389936] env[63021]: DEBUG oslo_vmware.api [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294140, 'name': ReconfigVM_Task, 'duration_secs': 0.298083} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.390230] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Reconfigured VM instance instance-0000005d to attach disk [datastore2] 9f822151-3307-4093-a972-8af128ea1892/9f822151-3307-4093-a972-8af128ea1892.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 998.390836] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ea99506d-482d-4d55-8971-e865c3d35234 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.397429] env[63021]: DEBUG oslo_vmware.api [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 998.397429] env[63021]: value = "task-1294141" [ 998.397429] env[63021]: _type = "Task" [ 998.397429] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.404634] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquiring lock "df856146-61ba-4de7-a5fd-d75263927163" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.404862] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "df856146-61ba-4de7-a5fd-d75263927163" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.405073] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquiring lock "df856146-61ba-4de7-a5fd-d75263927163-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.405269] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "df856146-61ba-4de7-a5fd-d75263927163-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.405437] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "df856146-61ba-4de7-a5fd-d75263927163-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.406971] env[63021]: DEBUG oslo_vmware.api [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294141, 'name': Rename_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.407625] env[63021]: INFO nova.compute.manager [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Terminating instance [ 998.409299] env[63021]: DEBUG nova.compute.manager [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 998.409494] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 998.410295] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d25590c5-5776-4153-bd9a-1351342432c4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.416181] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 998.416433] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-79059b20-8f17-4aa4-89e3-84b9f1d1a9d7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.421805] env[63021]: DEBUG oslo_vmware.api [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 998.421805] env[63021]: value = "task-1294142" [ 998.421805] env[63021]: _type = "Task" [ 998.421805] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.429377] env[63021]: DEBUG oslo_vmware.api [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294142, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.490480] env[63021]: DEBUG nova.compute.manager [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 998.578669] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.648938] env[63021]: DEBUG oslo_vmware.api [None req-b8fcad9f-92f2-40f7-bc7d-c79c2605d1cb tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294136, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.771935] env[63021]: DEBUG nova.scheduler.client.report [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 998.907518] env[63021]: DEBUG oslo_vmware.api [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294141, 'name': Rename_Task, 'duration_secs': 0.134246} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.907725] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 998.907978] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c881a7bb-3572-43fd-be2b-c864f32e0e79 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.914753] env[63021]: DEBUG oslo_vmware.api [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 998.914753] env[63021]: value = "task-1294143" [ 998.914753] env[63021]: _type = "Task" [ 998.914753] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.922859] env[63021]: DEBUG oslo_vmware.api [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294143, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.930043] env[63021]: DEBUG oslo_vmware.api [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294142, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.010410] env[63021]: DEBUG oslo_concurrency.lockutils [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.148721] env[63021]: DEBUG oslo_vmware.api [None req-b8fcad9f-92f2-40f7-bc7d-c79c2605d1cb tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294136, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.276990] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.261s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.277637] env[63021]: DEBUG nova.compute.manager [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 999.280146] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.702s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.282096] env[63021]: INFO nova.compute.claims [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 999.430667] env[63021]: DEBUG oslo_vmware.api [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294143, 'name': PowerOnVM_Task} progress is 94%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.436526] env[63021]: DEBUG oslo_vmware.api [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294142, 'name': PowerOffVM_Task, 'duration_secs': 0.555638} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.436526] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 999.436526] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 999.436526] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5fa0362c-f2b0-4f5d-92cc-c1503837bd14 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.519069] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 999.519342] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Deleting contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 999.519532] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Deleting the datastore file [datastore2] df856146-61ba-4de7-a5fd-d75263927163 {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 999.519859] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-27f61715-5862-4514-8fe2-aa2721bae50e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.525876] env[63021]: DEBUG oslo_vmware.api [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 999.525876] env[63021]: value = "task-1294145" [ 999.525876] env[63021]: _type = "Task" [ 999.525876] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.533853] env[63021]: DEBUG oslo_vmware.api [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294145, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.649729] env[63021]: DEBUG oslo_vmware.api [None req-b8fcad9f-92f2-40f7-bc7d-c79c2605d1cb tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294136, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.786493] env[63021]: DEBUG nova.compute.utils [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 999.791282] env[63021]: DEBUG nova.compute.manager [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 999.791642] env[63021]: DEBUG nova.network.neutron [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 999.846448] env[63021]: DEBUG nova.policy [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '31c5d292015043aa8bfc6647e7405ea3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5af87713ae6b477ca8a584aed287889d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 999.925312] env[63021]: DEBUG oslo_vmware.api [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294143, 'name': PowerOnVM_Task, 'duration_secs': 0.580375} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.925589] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 999.925800] env[63021]: INFO nova.compute.manager [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Took 7.25 seconds to spawn the instance on the hypervisor. [ 999.925978] env[63021]: DEBUG nova.compute.manager [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 999.926806] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c318c0e-c0e3-4175-88de-daa4a759589b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.036220] env[63021]: DEBUG oslo_vmware.api [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294145, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147728} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.036532] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1000.036724] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Deleted contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1000.036904] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1000.037087] env[63021]: INFO nova.compute.manager [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: df856146-61ba-4de7-a5fd-d75263927163] Took 1.63 seconds to destroy the instance on the hypervisor. [ 1000.037328] env[63021]: DEBUG oslo.service.loopingcall [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1000.037829] env[63021]: DEBUG nova.compute.manager [-] [instance: df856146-61ba-4de7-a5fd-d75263927163] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1000.037919] env[63021]: DEBUG nova.network.neutron [-] [instance: df856146-61ba-4de7-a5fd-d75263927163] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1000.089068] env[63021]: DEBUG nova.network.neutron [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Successfully created port: 371c655a-b09c-48dc-ad0a-da708503e86e {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1000.150276] env[63021]: DEBUG oslo_vmware.api [None req-b8fcad9f-92f2-40f7-bc7d-c79c2605d1cb tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294136, 'name': ReconfigVM_Task, 'duration_secs': 5.764493} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.150536] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b8fcad9f-92f2-40f7-bc7d-c79c2605d1cb tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Releasing lock "2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.150747] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b8fcad9f-92f2-40f7-bc7d-c79c2605d1cb tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Reconfigured VM to detach interface {{(pid=63021) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1000.292336] env[63021]: DEBUG nova.compute.manager [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1000.446332] env[63021]: INFO nova.compute.manager [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Took 14.08 seconds to build instance. [ 1000.499476] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-387b8b4c-3f44-4e01-b294-a8d9d9c617ea {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.507719] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e90ee14-79d8-4e0a-89ae-8a4f596d1cc5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.541058] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d5778ce-2af6-4d62-8f77-58ab6583b409 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.549258] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2f4f347-3f9d-4c52-bd92-5876023645fa {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.563653] env[63021]: DEBUG nova.compute.provider_tree [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1000.948896] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b9586113-0b67-499a-8b1b-9e83069b8dad tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "9f822151-3307-4093-a972-8af128ea1892" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.592s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.956236] env[63021]: DEBUG nova.compute.manager [req-584a4a45-cea3-4a3a-b2be-d27ab600d059 req-6be2c577-3e4c-423a-9b61-3d25ec189943 service nova] [instance: df856146-61ba-4de7-a5fd-d75263927163] Received event network-vif-deleted-a43dcfd2-397e-44fe-bd1c-b6116c25a28a {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1000.956236] env[63021]: INFO nova.compute.manager [req-584a4a45-cea3-4a3a-b2be-d27ab600d059 req-6be2c577-3e4c-423a-9b61-3d25ec189943 service nova] [instance: df856146-61ba-4de7-a5fd-d75263927163] Neutron deleted interface a43dcfd2-397e-44fe-bd1c-b6116c25a28a; detaching it from the instance and deleting it from the info cache [ 1000.956437] env[63021]: DEBUG nova.network.neutron [req-584a4a45-cea3-4a3a-b2be-d27ab600d059 req-6be2c577-3e4c-423a-9b61-3d25ec189943 service nova] [instance: df856146-61ba-4de7-a5fd-d75263927163] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.066701] env[63021]: DEBUG nova.scheduler.client.report [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1001.308957] env[63021]: DEBUG nova.compute.manager [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1001.334345] env[63021]: DEBUG nova.virt.hardware [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1001.334595] env[63021]: DEBUG nova.virt.hardware [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1001.334755] env[63021]: DEBUG nova.virt.hardware [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1001.334938] env[63021]: DEBUG nova.virt.hardware [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1001.335107] env[63021]: DEBUG nova.virt.hardware [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1001.335261] env[63021]: DEBUG nova.virt.hardware [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1001.335474] env[63021]: DEBUG nova.virt.hardware [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1001.335638] env[63021]: DEBUG nova.virt.hardware [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1001.335818] env[63021]: DEBUG nova.virt.hardware [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1001.335973] env[63021]: DEBUG nova.virt.hardware [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1001.336162] env[63021]: DEBUG nova.virt.hardware [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1001.337106] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-463b266a-9a72-4d74-b814-98f118b2d578 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.345227] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baf45350-fa13-4d99-afd2-aa9096dde311 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.429602] env[63021]: DEBUG nova.network.neutron [-] [instance: df856146-61ba-4de7-a5fd-d75263927163] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.459395] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e7887b9c-0252-4d7f-9922-cea26886ea6a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.470566] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0a11cc3-cfb1-43fc-a104-4e14eb148595 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.497463] env[63021]: DEBUG nova.compute.manager [req-584a4a45-cea3-4a3a-b2be-d27ab600d059 req-6be2c577-3e4c-423a-9b61-3d25ec189943 service nova] [instance: df856146-61ba-4de7-a5fd-d75263927163] Detach interface failed, port_id=a43dcfd2-397e-44fe-bd1c-b6116c25a28a, reason: Instance df856146-61ba-4de7-a5fd-d75263927163 could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1001.574709] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.294s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.575381] env[63021]: DEBUG nova.compute.manager [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1001.578467] env[63021]: DEBUG oslo_concurrency.lockutils [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.568s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.579553] env[63021]: INFO nova.compute.claims [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1001.591643] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b8fcad9f-92f2-40f7-bc7d-c79c2605d1cb tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "refresh_cache-2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.591814] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b8fcad9f-92f2-40f7-bc7d-c79c2605d1cb tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquired lock "refresh_cache-2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.591987] env[63021]: DEBUG nova.network.neutron [None req-b8fcad9f-92f2-40f7-bc7d-c79c2605d1cb tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1001.850605] env[63021]: DEBUG nova.compute.manager [req-bd4bdf03-0811-47f3-8a16-3610bf1cb46d req-73caad11-decf-4542-9fe0-1457cc7cb461 service nova] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Received event network-vif-plugged-371c655a-b09c-48dc-ad0a-da708503e86e {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1001.850841] env[63021]: DEBUG oslo_concurrency.lockutils [req-bd4bdf03-0811-47f3-8a16-3610bf1cb46d req-73caad11-decf-4542-9fe0-1457cc7cb461 service nova] Acquiring lock "67aed8eb-d08c-47cc-8558-5c5a0f058d6e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.851059] env[63021]: DEBUG oslo_concurrency.lockutils [req-bd4bdf03-0811-47f3-8a16-3610bf1cb46d req-73caad11-decf-4542-9fe0-1457cc7cb461 service nova] Lock "67aed8eb-d08c-47cc-8558-5c5a0f058d6e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.851237] env[63021]: DEBUG oslo_concurrency.lockutils [req-bd4bdf03-0811-47f3-8a16-3610bf1cb46d req-73caad11-decf-4542-9fe0-1457cc7cb461 service nova] Lock "67aed8eb-d08c-47cc-8558-5c5a0f058d6e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.851460] env[63021]: DEBUG nova.compute.manager [req-bd4bdf03-0811-47f3-8a16-3610bf1cb46d req-73caad11-decf-4542-9fe0-1457cc7cb461 service nova] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] No waiting events found dispatching network-vif-plugged-371c655a-b09c-48dc-ad0a-da708503e86e {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1001.851573] env[63021]: WARNING nova.compute.manager [req-bd4bdf03-0811-47f3-8a16-3610bf1cb46d req-73caad11-decf-4542-9fe0-1457cc7cb461 service nova] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Received unexpected event network-vif-plugged-371c655a-b09c-48dc-ad0a-da708503e86e for instance with vm_state building and task_state spawning. [ 1001.909275] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "9f822151-3307-4093-a972-8af128ea1892" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.909538] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "9f822151-3307-4093-a972-8af128ea1892" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.909754] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "9f822151-3307-4093-a972-8af128ea1892-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.910058] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "9f822151-3307-4093-a972-8af128ea1892-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.910255] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "9f822151-3307-4093-a972-8af128ea1892-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.912549] env[63021]: INFO nova.compute.manager [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Terminating instance [ 1001.914189] env[63021]: DEBUG nova.compute.manager [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1001.914390] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1001.915220] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b26f1382-1e33-41c9-9021-e57bfcfae582 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.923451] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1001.924023] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-df61b0fa-a8f6-4ab9-bf3f-d0e1f96c498c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.933927] env[63021]: DEBUG nova.network.neutron [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Successfully updated port: 371c655a-b09c-48dc-ad0a-da708503e86e {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1001.939023] env[63021]: INFO nova.compute.manager [-] [instance: df856146-61ba-4de7-a5fd-d75263927163] Took 1.90 seconds to deallocate network for instance. [ 1001.941869] env[63021]: DEBUG oslo_vmware.api [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 1001.941869] env[63021]: value = "task-1294146" [ 1001.941869] env[63021]: _type = "Task" [ 1001.941869] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.957855] env[63021]: DEBUG oslo_vmware.api [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294146, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.083917] env[63021]: DEBUG nova.compute.utils [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1002.088115] env[63021]: DEBUG nova.compute.manager [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1002.088401] env[63021]: DEBUG nova.network.neutron [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1002.146088] env[63021]: DEBUG nova.policy [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ba16721213dd49a3b4e4990d7d319d1f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3c8959d2f1fe4706b7407c11e128b773', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 1002.401407] env[63021]: DEBUG nova.network.neutron [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Successfully created port: 2df835dc-fa26-45ab-9d30-58f85c9e3331 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1002.436710] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "refresh_cache-67aed8eb-d08c-47cc-8558-5c5a0f058d6e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.436920] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquired lock "refresh_cache-67aed8eb-d08c-47cc-8558-5c5a0f058d6e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.437423] env[63021]: DEBUG nova.network.neutron [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1002.452566] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.459687] env[63021]: DEBUG oslo_vmware.api [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294146, 'name': PowerOffVM_Task, 'duration_secs': 0.203841} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.460779] env[63021]: INFO nova.network.neutron [None req-b8fcad9f-92f2-40f7-bc7d-c79c2605d1cb tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Port 01e26b49-58a2-411d-b025-b1ca941026c4 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1002.461171] env[63021]: DEBUG nova.network.neutron [None req-b8fcad9f-92f2-40f7-bc7d-c79c2605d1cb tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Updating instance_info_cache with network_info: [{"id": "eab7f4fe-4832-40e4-aed2-0e931d91a4d7", "address": "fa:16:3e:9c:24:88", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeab7f4fe-48", "ovs_interfaceid": "eab7f4fe-4832-40e4-aed2-0e931d91a4d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.462659] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1002.462860] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1002.463176] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7b51097e-8ae0-477e-aa9b-ffc4a9755559 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.528439] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1002.528741] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Deleting contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1002.528941] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Deleting the datastore file [datastore2] 9f822151-3307-4093-a972-8af128ea1892 {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1002.529261] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-87d54263-c90f-4180-ae9f-605aca378e75 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.537577] env[63021]: DEBUG oslo_vmware.api [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for the task: (returnval){ [ 1002.537577] env[63021]: value = "task-1294148" [ 1002.537577] env[63021]: _type = "Task" [ 1002.537577] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.548813] env[63021]: DEBUG oslo_vmware.api [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294148, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.588711] env[63021]: DEBUG nova.compute.manager [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1002.783707] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd30856a-7c41-4f18-b4e3-250919e1bd1d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.794381] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8c3c521-b0ad-439d-abb2-ba98b10abb9d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.825420] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd623f06-b715-4de2-a54c-89e4c5b9c525 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.833101] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9f2424b-7881-4252-90f7-473945afb867 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.847641] env[63021]: DEBUG nova.compute.provider_tree [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1002.967026] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b8fcad9f-92f2-40f7-bc7d-c79c2605d1cb tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Releasing lock "refresh_cache-2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.976012] env[63021]: DEBUG nova.network.neutron [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1002.982126] env[63021]: DEBUG nova.compute.manager [req-4c2188db-a81e-436d-a17f-2bf2b17ced51 req-bfcad304-e237-45d2-b63f-39a5f805acc9 service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Received event network-changed-eab7f4fe-4832-40e4-aed2-0e931d91a4d7 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1002.982447] env[63021]: DEBUG nova.compute.manager [req-4c2188db-a81e-436d-a17f-2bf2b17ced51 req-bfcad304-e237-45d2-b63f-39a5f805acc9 service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Refreshing instance network info cache due to event network-changed-eab7f4fe-4832-40e4-aed2-0e931d91a4d7. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1002.982534] env[63021]: DEBUG oslo_concurrency.lockutils [req-4c2188db-a81e-436d-a17f-2bf2b17ced51 req-bfcad304-e237-45d2-b63f-39a5f805acc9 service nova] Acquiring lock "refresh_cache-2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.982675] env[63021]: DEBUG oslo_concurrency.lockutils [req-4c2188db-a81e-436d-a17f-2bf2b17ced51 req-bfcad304-e237-45d2-b63f-39a5f805acc9 service nova] Acquired lock "refresh_cache-2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.982844] env[63021]: DEBUG nova.network.neutron [req-4c2188db-a81e-436d-a17f-2bf2b17ced51 req-bfcad304-e237-45d2-b63f-39a5f805acc9 service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Refreshing network info cache for port eab7f4fe-4832-40e4-aed2-0e931d91a4d7 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1003.050946] env[63021]: DEBUG oslo_vmware.api [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Task: {'id': task-1294148, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.170492} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.051452] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1003.051660] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Deleted contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1003.051865] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1003.052057] env[63021]: INFO nova.compute.manager [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] [instance: 9f822151-3307-4093-a972-8af128ea1892] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1003.052333] env[63021]: DEBUG oslo.service.loopingcall [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1003.052554] env[63021]: DEBUG nova.compute.manager [-] [instance: 9f822151-3307-4093-a972-8af128ea1892] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1003.052651] env[63021]: DEBUG nova.network.neutron [-] [instance: 9f822151-3307-4093-a972-8af128ea1892] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1003.134374] env[63021]: DEBUG nova.network.neutron [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Updating instance_info_cache with network_info: [{"id": "371c655a-b09c-48dc-ad0a-da708503e86e", "address": "fa:16:3e:6b:ae:03", "network": {"id": "9502e364-60e1-47ef-a1f7-f3e3ff45cf69", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-164717784-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5af87713ae6b477ca8a584aed287889d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap371c655a-b0", "ovs_interfaceid": "371c655a-b09c-48dc-ad0a-da708503e86e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.213216] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "interface-6155c098-e4a6-47e6-b343-4a77ca90eb2e-01e26b49-58a2-411d-b025-b1ca941026c4" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.213481] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "interface-6155c098-e4a6-47e6-b343-4a77ca90eb2e-01e26b49-58a2-411d-b025-b1ca941026c4" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.213851] env[63021]: DEBUG nova.objects.instance [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lazy-loading 'flavor' on Instance uuid 6155c098-e4a6-47e6-b343-4a77ca90eb2e {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1003.349266] env[63021]: DEBUG nova.scheduler.client.report [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1003.470726] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b8fcad9f-92f2-40f7-bc7d-c79c2605d1cb tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "interface-2d647028-306a-41d5-96f0-2e82a98ab56d-01e26b49-58a2-411d-b025-b1ca941026c4" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.910s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.601018] env[63021]: DEBUG nova.compute.manager [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1003.627137] env[63021]: DEBUG nova.virt.hardware [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1003.627401] env[63021]: DEBUG nova.virt.hardware [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1003.627560] env[63021]: DEBUG nova.virt.hardware [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1003.627744] env[63021]: DEBUG nova.virt.hardware [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1003.627893] env[63021]: DEBUG nova.virt.hardware [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1003.628052] env[63021]: DEBUG nova.virt.hardware [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1003.628262] env[63021]: DEBUG nova.virt.hardware [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1003.628439] env[63021]: DEBUG nova.virt.hardware [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1003.628634] env[63021]: DEBUG nova.virt.hardware [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1003.628803] env[63021]: DEBUG nova.virt.hardware [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1003.629007] env[63021]: DEBUG nova.virt.hardware [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1003.629841] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-561a320b-6540-450c-91be-41b06f1a899a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.636572] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Releasing lock "refresh_cache-67aed8eb-d08c-47cc-8558-5c5a0f058d6e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.636844] env[63021]: DEBUG nova.compute.manager [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Instance network_info: |[{"id": "371c655a-b09c-48dc-ad0a-da708503e86e", "address": "fa:16:3e:6b:ae:03", "network": {"id": "9502e364-60e1-47ef-a1f7-f3e3ff45cf69", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-164717784-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5af87713ae6b477ca8a584aed287889d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap371c655a-b0", "ovs_interfaceid": "371c655a-b09c-48dc-ad0a-da708503e86e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1003.637280] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6b:ae:03', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '371c655a-b09c-48dc-ad0a-da708503e86e', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1003.644704] env[63021]: DEBUG oslo.service.loopingcall [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1003.645850] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1306fd5d-8f60-4f4c-b7cd-9a638a2d958d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.652315] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1003.652538] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-106ed8e0-09c4-4c13-bb84-7ae867297637 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.679180] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1003.679180] env[63021]: value = "task-1294149" [ 1003.679180] env[63021]: _type = "Task" [ 1003.679180] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.687961] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294149, 'name': CreateVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.720112] env[63021]: DEBUG nova.network.neutron [req-4c2188db-a81e-436d-a17f-2bf2b17ced51 req-bfcad304-e237-45d2-b63f-39a5f805acc9 service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Updated VIF entry in instance network info cache for port eab7f4fe-4832-40e4-aed2-0e931d91a4d7. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1003.720466] env[63021]: DEBUG nova.network.neutron [req-4c2188db-a81e-436d-a17f-2bf2b17ced51 req-bfcad304-e237-45d2-b63f-39a5f805acc9 service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Updating instance_info_cache with network_info: [{"id": "eab7f4fe-4832-40e4-aed2-0e931d91a4d7", "address": "fa:16:3e:9c:24:88", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeab7f4fe-48", "ovs_interfaceid": "eab7f4fe-4832-40e4-aed2-0e931d91a4d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.786762] env[63021]: DEBUG nova.network.neutron [-] [instance: 9f822151-3307-4093-a972-8af128ea1892] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.855020] env[63021]: DEBUG oslo_concurrency.lockutils [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.275s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.855020] env[63021]: DEBUG nova.compute.manager [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1003.859135] env[63021]: DEBUG nova.objects.instance [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lazy-loading 'pci_requests' on Instance uuid 6155c098-e4a6-47e6-b343-4a77ca90eb2e {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1003.860133] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.408s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.860856] env[63021]: DEBUG nova.objects.instance [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lazy-loading 'resources' on Instance uuid df856146-61ba-4de7-a5fd-d75263927163 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1003.957088] env[63021]: DEBUG nova.compute.manager [req-ab829ab7-639e-492c-8250-445adfa9b97d req-7a4905ce-07fe-4482-a4c3-dc592d1118d9 service nova] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Received event network-changed-371c655a-b09c-48dc-ad0a-da708503e86e {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1003.957088] env[63021]: DEBUG nova.compute.manager [req-ab829ab7-639e-492c-8250-445adfa9b97d req-7a4905ce-07fe-4482-a4c3-dc592d1118d9 service nova] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Refreshing instance network info cache due to event network-changed-371c655a-b09c-48dc-ad0a-da708503e86e. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1003.957088] env[63021]: DEBUG oslo_concurrency.lockutils [req-ab829ab7-639e-492c-8250-445adfa9b97d req-7a4905ce-07fe-4482-a4c3-dc592d1118d9 service nova] Acquiring lock "refresh_cache-67aed8eb-d08c-47cc-8558-5c5a0f058d6e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.957088] env[63021]: DEBUG oslo_concurrency.lockutils [req-ab829ab7-639e-492c-8250-445adfa9b97d req-7a4905ce-07fe-4482-a4c3-dc592d1118d9 service nova] Acquired lock "refresh_cache-67aed8eb-d08c-47cc-8558-5c5a0f058d6e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.957088] env[63021]: DEBUG nova.network.neutron [req-ab829ab7-639e-492c-8250-445adfa9b97d req-7a4905ce-07fe-4482-a4c3-dc592d1118d9 service nova] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Refreshing network info cache for port 371c655a-b09c-48dc-ad0a-da708503e86e {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1003.957663] env[63021]: DEBUG nova.network.neutron [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Successfully updated port: 2df835dc-fa26-45ab-9d30-58f85c9e3331 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1004.189542] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294149, 'name': CreateVM_Task, 'duration_secs': 0.508937} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.189846] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1004.190441] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.190618] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.190967] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1004.191256] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-09d86a85-0752-4df3-b733-6b048cdad207 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.196023] env[63021]: DEBUG oslo_vmware.api [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1004.196023] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52dd8f58-8352-d820-efdd-92bdff8fa537" [ 1004.196023] env[63021]: _type = "Task" [ 1004.196023] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.203801] env[63021]: DEBUG oslo_vmware.api [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52dd8f58-8352-d820-efdd-92bdff8fa537, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.222935] env[63021]: DEBUG oslo_concurrency.lockutils [req-4c2188db-a81e-436d-a17f-2bf2b17ced51 req-bfcad304-e237-45d2-b63f-39a5f805acc9 service nova] Releasing lock "refresh_cache-2d647028-306a-41d5-96f0-2e82a98ab56d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.223269] env[63021]: DEBUG nova.compute.manager [req-4c2188db-a81e-436d-a17f-2bf2b17ced51 req-bfcad304-e237-45d2-b63f-39a5f805acc9 service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Received event network-changed-3752044a-9145-469d-8edb-6255d7d0f928 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1004.223444] env[63021]: DEBUG nova.compute.manager [req-4c2188db-a81e-436d-a17f-2bf2b17ced51 req-bfcad304-e237-45d2-b63f-39a5f805acc9 service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Refreshing instance network info cache due to event network-changed-3752044a-9145-469d-8edb-6255d7d0f928. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1004.223657] env[63021]: DEBUG oslo_concurrency.lockutils [req-4c2188db-a81e-436d-a17f-2bf2b17ced51 req-bfcad304-e237-45d2-b63f-39a5f805acc9 service nova] Acquiring lock "refresh_cache-6155c098-e4a6-47e6-b343-4a77ca90eb2e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.223800] env[63021]: DEBUG oslo_concurrency.lockutils [req-4c2188db-a81e-436d-a17f-2bf2b17ced51 req-bfcad304-e237-45d2-b63f-39a5f805acc9 service nova] Acquired lock "refresh_cache-6155c098-e4a6-47e6-b343-4a77ca90eb2e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.223964] env[63021]: DEBUG nova.network.neutron [req-4c2188db-a81e-436d-a17f-2bf2b17ced51 req-bfcad304-e237-45d2-b63f-39a5f805acc9 service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Refreshing network info cache for port 3752044a-9145-469d-8edb-6255d7d0f928 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1004.289836] env[63021]: INFO nova.compute.manager [-] [instance: 9f822151-3307-4093-a972-8af128ea1892] Took 1.24 seconds to deallocate network for instance. [ 1004.362879] env[63021]: DEBUG nova.compute.utils [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1004.364364] env[63021]: DEBUG nova.objects.base [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Object Instance<6155c098-e4a6-47e6-b343-4a77ca90eb2e> lazy-loaded attributes: flavor,pci_requests {{(pid=63021) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1004.364651] env[63021]: DEBUG nova.network.neutron [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1004.366777] env[63021]: DEBUG nova.compute.manager [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1004.366965] env[63021]: DEBUG nova.network.neutron [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1004.413429] env[63021]: DEBUG nova.policy [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b427f7d1317341f38754a2b324f4f5aa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e238d5c1668540d2a9fffd3fd832d9b8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 1004.433754] env[63021]: DEBUG nova.policy [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6ea19453dbdb4251a93dde2576aeff64', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '58693d1e991249f28914a39761cc2627', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 1004.460459] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "refresh_cache-023c0891-2b08-48bb-9754-408ce7d04c9c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.460459] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquired lock "refresh_cache-023c0891-2b08-48bb-9754-408ce7d04c9c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.460614] env[63021]: DEBUG nova.network.neutron [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1004.524187] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13612a7a-2b1b-4866-9030-17cfa64747f5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.531994] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59524908-1a47-405c-a334-fb7678be3cdd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.564767] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bca7d8a0-55f8-4597-977c-68ba0ba6ccbb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.572136] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-900796e4-6844-48d7-b497-4048f708838b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.586346] env[63021]: DEBUG nova.compute.provider_tree [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1004.648443] env[63021]: DEBUG nova.network.neutron [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Successfully created port: 4a29562f-684a-4738-98b0-f1281b0bc4c3 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1004.707421] env[63021]: DEBUG oslo_vmware.api [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52dd8f58-8352-d820-efdd-92bdff8fa537, 'name': SearchDatastore_Task, 'duration_secs': 0.008342} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.707776] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.708021] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1004.708259] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.708407] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.708588] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1004.708879] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-01be201f-097e-4926-a4d9-5fbbe8f00541 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.717051] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1004.717248] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1004.717940] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-044a0de3-5761-42ea-92ca-5dd195ec4125 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.722908] env[63021]: DEBUG oslo_vmware.api [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1004.722908] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52d3733c-a3c3-1c9a-10b1-662f0dca2fe4" [ 1004.722908] env[63021]: _type = "Task" [ 1004.722908] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.735141] env[63021]: DEBUG oslo_vmware.api [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52d3733c-a3c3-1c9a-10b1-662f0dca2fe4, 'name': SearchDatastore_Task, 'duration_secs': 0.007814} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.735876] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55b16e98-d530-474f-9c85-a0f022905649 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.740768] env[63021]: DEBUG oslo_vmware.api [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1004.740768] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52f3f716-c24e-58cf-f54a-2a50591536bb" [ 1004.740768] env[63021]: _type = "Task" [ 1004.740768] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.748047] env[63021]: DEBUG oslo_vmware.api [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52f3f716-c24e-58cf-f54a-2a50591536bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.796306] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.817748] env[63021]: DEBUG nova.network.neutron [req-ab829ab7-639e-492c-8250-445adfa9b97d req-7a4905ce-07fe-4482-a4c3-dc592d1118d9 service nova] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Updated VIF entry in instance network info cache for port 371c655a-b09c-48dc-ad0a-da708503e86e. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1004.818070] env[63021]: DEBUG nova.network.neutron [req-ab829ab7-639e-492c-8250-445adfa9b97d req-7a4905ce-07fe-4482-a4c3-dc592d1118d9 service nova] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Updating instance_info_cache with network_info: [{"id": "371c655a-b09c-48dc-ad0a-da708503e86e", "address": "fa:16:3e:6b:ae:03", "network": {"id": "9502e364-60e1-47ef-a1f7-f3e3ff45cf69", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-164717784-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5af87713ae6b477ca8a584aed287889d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap371c655a-b0", "ovs_interfaceid": "371c655a-b09c-48dc-ad0a-da708503e86e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.869770] env[63021]: DEBUG nova.compute.manager [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1005.000313] env[63021]: DEBUG nova.network.neutron [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1005.047281] env[63021]: DEBUG nova.network.neutron [req-4c2188db-a81e-436d-a17f-2bf2b17ced51 req-bfcad304-e237-45d2-b63f-39a5f805acc9 service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Updated VIF entry in instance network info cache for port 3752044a-9145-469d-8edb-6255d7d0f928. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1005.047671] env[63021]: DEBUG nova.network.neutron [req-4c2188db-a81e-436d-a17f-2bf2b17ced51 req-bfcad304-e237-45d2-b63f-39a5f805acc9 service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Updating instance_info_cache with network_info: [{"id": "3752044a-9145-469d-8edb-6255d7d0f928", "address": "fa:16:3e:fe:57:1f", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3752044a-91", "ovs_interfaceid": "3752044a-9145-469d-8edb-6255d7d0f928", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.090908] env[63021]: DEBUG nova.scheduler.client.report [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1005.150644] env[63021]: DEBUG nova.network.neutron [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Updating instance_info_cache with network_info: [{"id": "2df835dc-fa26-45ab-9d30-58f85c9e3331", "address": "fa:16:3e:d9:6c:e7", "network": {"id": "0c0dd1a7-6687-45c5-82b1-33f34bdbb848", "bridge": "br-int", "label": "tempest-ServersTestJSON-523240467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c8959d2f1fe4706b7407c11e128b773", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffc811e4-0e4e-4748-8c7e-b3f14ccbd42d", "external-id": "nsx-vlan-transportzone-404", "segmentation_id": 404, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2df835dc-fa", "ovs_interfaceid": "2df835dc-fa26-45ab-9d30-58f85c9e3331", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.250629] env[63021]: DEBUG oslo_vmware.api [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52f3f716-c24e-58cf-f54a-2a50591536bb, 'name': SearchDatastore_Task, 'duration_secs': 0.00803} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.250953] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.251176] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 67aed8eb-d08c-47cc-8558-5c5a0f058d6e/67aed8eb-d08c-47cc-8558-5c5a0f058d6e.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1005.251440] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2d4e3de5-f7cb-4f11-9e33-e89346bd4afc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.258519] env[63021]: DEBUG oslo_vmware.api [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1005.258519] env[63021]: value = "task-1294150" [ 1005.258519] env[63021]: _type = "Task" [ 1005.258519] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.267331] env[63021]: DEBUG oslo_vmware.api [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294150, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.320763] env[63021]: DEBUG oslo_concurrency.lockutils [req-ab829ab7-639e-492c-8250-445adfa9b97d req-7a4905ce-07fe-4482-a4c3-dc592d1118d9 service nova] Releasing lock "refresh_cache-67aed8eb-d08c-47cc-8558-5c5a0f058d6e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.321151] env[63021]: DEBUG nova.compute.manager [req-ab829ab7-639e-492c-8250-445adfa9b97d req-7a4905ce-07fe-4482-a4c3-dc592d1118d9 service nova] [instance: 9f822151-3307-4093-a972-8af128ea1892] Received event network-vif-deleted-e5cc5b77-3f1d-444f-bb24-2d7f98b67a8b {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1005.321864] env[63021]: DEBUG nova.compute.manager [req-ab829ab7-639e-492c-8250-445adfa9b97d req-7a4905ce-07fe-4482-a4c3-dc592d1118d9 service nova] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Received event network-vif-plugged-2df835dc-fa26-45ab-9d30-58f85c9e3331 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1005.321864] env[63021]: DEBUG oslo_concurrency.lockutils [req-ab829ab7-639e-492c-8250-445adfa9b97d req-7a4905ce-07fe-4482-a4c3-dc592d1118d9 service nova] Acquiring lock "023c0891-2b08-48bb-9754-408ce7d04c9c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.322033] env[63021]: DEBUG oslo_concurrency.lockutils [req-ab829ab7-639e-492c-8250-445adfa9b97d req-7a4905ce-07fe-4482-a4c3-dc592d1118d9 service nova] Lock "023c0891-2b08-48bb-9754-408ce7d04c9c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.322123] env[63021]: DEBUG oslo_concurrency.lockutils [req-ab829ab7-639e-492c-8250-445adfa9b97d req-7a4905ce-07fe-4482-a4c3-dc592d1118d9 service nova] Lock "023c0891-2b08-48bb-9754-408ce7d04c9c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.322297] env[63021]: DEBUG nova.compute.manager [req-ab829ab7-639e-492c-8250-445adfa9b97d req-7a4905ce-07fe-4482-a4c3-dc592d1118d9 service nova] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] No waiting events found dispatching network-vif-plugged-2df835dc-fa26-45ab-9d30-58f85c9e3331 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1005.322468] env[63021]: WARNING nova.compute.manager [req-ab829ab7-639e-492c-8250-445adfa9b97d req-7a4905ce-07fe-4482-a4c3-dc592d1118d9 service nova] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Received unexpected event network-vif-plugged-2df835dc-fa26-45ab-9d30-58f85c9e3331 for instance with vm_state building and task_state spawning. [ 1005.550471] env[63021]: DEBUG oslo_concurrency.lockutils [req-4c2188db-a81e-436d-a17f-2bf2b17ced51 req-bfcad304-e237-45d2-b63f-39a5f805acc9 service nova] Releasing lock "refresh_cache-6155c098-e4a6-47e6-b343-4a77ca90eb2e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.596786] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.736s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.599215] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.803s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.599455] env[63021]: DEBUG nova.objects.instance [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lazy-loading 'resources' on Instance uuid 9f822151-3307-4093-a972-8af128ea1892 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.620385] env[63021]: INFO nova.scheduler.client.report [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Deleted allocations for instance df856146-61ba-4de7-a5fd-d75263927163 [ 1005.653185] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Releasing lock "refresh_cache-023c0891-2b08-48bb-9754-408ce7d04c9c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.653464] env[63021]: DEBUG nova.compute.manager [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Instance network_info: |[{"id": "2df835dc-fa26-45ab-9d30-58f85c9e3331", "address": "fa:16:3e:d9:6c:e7", "network": {"id": "0c0dd1a7-6687-45c5-82b1-33f34bdbb848", "bridge": "br-int", "label": "tempest-ServersTestJSON-523240467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c8959d2f1fe4706b7407c11e128b773", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffc811e4-0e4e-4748-8c7e-b3f14ccbd42d", "external-id": "nsx-vlan-transportzone-404", "segmentation_id": 404, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2df835dc-fa", "ovs_interfaceid": "2df835dc-fa26-45ab-9d30-58f85c9e3331", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1005.653934] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d9:6c:e7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ffc811e4-0e4e-4748-8c7e-b3f14ccbd42d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2df835dc-fa26-45ab-9d30-58f85c9e3331', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1005.661892] env[63021]: DEBUG oslo.service.loopingcall [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1005.662591] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1005.662591] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8e570f89-9519-4ca2-9ce4-66bc1445025a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.683352] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1005.683352] env[63021]: value = "task-1294151" [ 1005.683352] env[63021]: _type = "Task" [ 1005.683352] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.694493] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294151, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.768455] env[63021]: DEBUG oslo_vmware.api [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294150, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.453419} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.768792] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 67aed8eb-d08c-47cc-8558-5c5a0f058d6e/67aed8eb-d08c-47cc-8558-5c5a0f058d6e.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1005.769024] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1005.769287] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9595b0b5-4035-4462-9842-25f220bbe216 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.775017] env[63021]: DEBUG oslo_vmware.api [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1005.775017] env[63021]: value = "task-1294152" [ 1005.775017] env[63021]: _type = "Task" [ 1005.775017] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.782131] env[63021]: DEBUG oslo_vmware.api [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294152, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.879118] env[63021]: DEBUG nova.compute.manager [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1005.906122] env[63021]: DEBUG nova.virt.hardware [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1005.906474] env[63021]: DEBUG nova.virt.hardware [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1005.906548] env[63021]: DEBUG nova.virt.hardware [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1005.906763] env[63021]: DEBUG nova.virt.hardware [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1005.906979] env[63021]: DEBUG nova.virt.hardware [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1005.907072] env[63021]: DEBUG nova.virt.hardware [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1005.907294] env[63021]: DEBUG nova.virt.hardware [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1005.907485] env[63021]: DEBUG nova.virt.hardware [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1005.907773] env[63021]: DEBUG nova.virt.hardware [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1005.908046] env[63021]: DEBUG nova.virt.hardware [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1005.908097] env[63021]: DEBUG nova.virt.hardware [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1005.909010] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a16b268-3fbb-48b2-be17-01f385b14072 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.917685] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c800c1c-3fd0-4040-b657-4021ffbbe323 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.978681] env[63021]: DEBUG nova.compute.manager [req-2a271628-3230-421f-bd59-2937d90c1f30 req-ad5bb27a-91db-40e4-a890-2fbdcebba064 service nova] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Received event network-changed-2df835dc-fa26-45ab-9d30-58f85c9e3331 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1005.978909] env[63021]: DEBUG nova.compute.manager [req-2a271628-3230-421f-bd59-2937d90c1f30 req-ad5bb27a-91db-40e4-a890-2fbdcebba064 service nova] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Refreshing instance network info cache due to event network-changed-2df835dc-fa26-45ab-9d30-58f85c9e3331. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1005.979189] env[63021]: DEBUG oslo_concurrency.lockutils [req-2a271628-3230-421f-bd59-2937d90c1f30 req-ad5bb27a-91db-40e4-a890-2fbdcebba064 service nova] Acquiring lock "refresh_cache-023c0891-2b08-48bb-9754-408ce7d04c9c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.979342] env[63021]: DEBUG oslo_concurrency.lockutils [req-2a271628-3230-421f-bd59-2937d90c1f30 req-ad5bb27a-91db-40e4-a890-2fbdcebba064 service nova] Acquired lock "refresh_cache-023c0891-2b08-48bb-9754-408ce7d04c9c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.979510] env[63021]: DEBUG nova.network.neutron [req-2a271628-3230-421f-bd59-2937d90c1f30 req-ad5bb27a-91db-40e4-a890-2fbdcebba064 service nova] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Refreshing network info cache for port 2df835dc-fa26-45ab-9d30-58f85c9e3331 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1006.054426] env[63021]: DEBUG nova.compute.manager [req-2be57609-3bef-4618-89cf-a6c4ff340944 req-60171fd0-ee13-45b0-9250-b677824070b8 service nova] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Received event network-vif-plugged-4a29562f-684a-4738-98b0-f1281b0bc4c3 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1006.054724] env[63021]: DEBUG oslo_concurrency.lockutils [req-2be57609-3bef-4618-89cf-a6c4ff340944 req-60171fd0-ee13-45b0-9250-b677824070b8 service nova] Acquiring lock "07510760-13ff-4797-86ca-86b384aacdfa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.054985] env[63021]: DEBUG oslo_concurrency.lockutils [req-2be57609-3bef-4618-89cf-a6c4ff340944 req-60171fd0-ee13-45b0-9250-b677824070b8 service nova] Lock "07510760-13ff-4797-86ca-86b384aacdfa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.055057] env[63021]: DEBUG oslo_concurrency.lockutils [req-2be57609-3bef-4618-89cf-a6c4ff340944 req-60171fd0-ee13-45b0-9250-b677824070b8 service nova] Lock "07510760-13ff-4797-86ca-86b384aacdfa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.055228] env[63021]: DEBUG nova.compute.manager [req-2be57609-3bef-4618-89cf-a6c4ff340944 req-60171fd0-ee13-45b0-9250-b677824070b8 service nova] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] No waiting events found dispatching network-vif-plugged-4a29562f-684a-4738-98b0-f1281b0bc4c3 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1006.055352] env[63021]: WARNING nova.compute.manager [req-2be57609-3bef-4618-89cf-a6c4ff340944 req-60171fd0-ee13-45b0-9250-b677824070b8 service nova] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Received unexpected event network-vif-plugged-4a29562f-684a-4738-98b0-f1281b0bc4c3 for instance with vm_state building and task_state spawning. [ 1006.127649] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bd1f967e-6d0b-4c48-af4b-6a8c151a7a9f tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "df856146-61ba-4de7-a5fd-d75263927163" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.723s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.174064] env[63021]: DEBUG nova.network.neutron [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Successfully updated port: 01e26b49-58a2-411d-b025-b1ca941026c4 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1006.198119] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294151, 'name': CreateVM_Task, 'duration_secs': 0.344663} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.198995] env[63021]: DEBUG nova.network.neutron [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Successfully updated port: 4a29562f-684a-4738-98b0-f1281b0bc4c3 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1006.200472] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1006.201227] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.201396] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.201759] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1006.202431] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d6bcc1a6-10d7-4e7d-be97-20f1177acd0b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.211971] env[63021]: DEBUG oslo_vmware.api [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1006.211971] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]528e32a9-e3fe-9f32-5653-a2c4a7119bc9" [ 1006.211971] env[63021]: _type = "Task" [ 1006.211971] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.229710] env[63021]: DEBUG oslo_vmware.api [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]528e32a9-e3fe-9f32-5653-a2c4a7119bc9, 'name': SearchDatastore_Task, 'duration_secs': 0.0097} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.229710] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.229710] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1006.229710] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.229710] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.229710] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1006.229710] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ab48ac40-f3b8-45ea-8e1e-25f407974d60 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.237430] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1006.237430] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1006.237691] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99c6442e-11bb-4c24-8d4c-8301ff86b1f2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.246836] env[63021]: DEBUG oslo_vmware.api [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1006.246836] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]523370e3-38e7-88bc-44ad-e57b84664672" [ 1006.246836] env[63021]: _type = "Task" [ 1006.246836] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.254787] env[63021]: DEBUG oslo_vmware.api [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]523370e3-38e7-88bc-44ad-e57b84664672, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.285764] env[63021]: DEBUG oslo_vmware.api [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294152, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061703} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.286116] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1006.287083] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c336a618-7872-4688-a97f-1bfc00529785 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.310651] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] 67aed8eb-d08c-47cc-8558-5c5a0f058d6e/67aed8eb-d08c-47cc-8558-5c5a0f058d6e.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1006.311032] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-be0b48ef-d000-4efb-8e77-a09d35f30b78 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.333706] env[63021]: DEBUG oslo_vmware.api [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1006.333706] env[63021]: value = "task-1294153" [ 1006.333706] env[63021]: _type = "Task" [ 1006.333706] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.338791] env[63021]: DEBUG oslo_vmware.api [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294153, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.414649] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c084fd9a-35df-4e11-b038-0fe58454f169 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.422132] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64569be6-56e8-457c-8786-6df404d0a21a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.455059] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f8f1fe2-7092-4a99-bb40-c384105b3bd7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.462299] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58ccf03c-9963-4a56-ac1c-09b9a92ae79e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.477454] env[63021]: DEBUG nova.compute.provider_tree [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1006.681064] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "refresh_cache-6155c098-e4a6-47e6-b343-4a77ca90eb2e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.681064] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquired lock "refresh_cache-6155c098-e4a6-47e6-b343-4a77ca90eb2e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.681064] env[63021]: DEBUG nova.network.neutron [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1006.701292] env[63021]: DEBUG oslo_concurrency.lockutils [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "refresh_cache-07510760-13ff-4797-86ca-86b384aacdfa" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.701434] env[63021]: DEBUG oslo_concurrency.lockutils [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquired lock "refresh_cache-07510760-13ff-4797-86ca-86b384aacdfa" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.701576] env[63021]: DEBUG nova.network.neutron [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1006.756943] env[63021]: DEBUG oslo_vmware.api [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]523370e3-38e7-88bc-44ad-e57b84664672, 'name': SearchDatastore_Task, 'duration_secs': 0.009528} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.757749] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f43ef494-884d-42e4-92eb-9f1ccdbe5d54 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.762634] env[63021]: DEBUG oslo_vmware.api [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1006.762634] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]529ec01f-324d-8559-3382-57bd2badf8c3" [ 1006.762634] env[63021]: _type = "Task" [ 1006.762634] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.770280] env[63021]: DEBUG oslo_vmware.api [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]529ec01f-324d-8559-3382-57bd2badf8c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.775646] env[63021]: DEBUG nova.network.neutron [req-2a271628-3230-421f-bd59-2937d90c1f30 req-ad5bb27a-91db-40e4-a890-2fbdcebba064 service nova] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Updated VIF entry in instance network info cache for port 2df835dc-fa26-45ab-9d30-58f85c9e3331. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1006.775981] env[63021]: DEBUG nova.network.neutron [req-2a271628-3230-421f-bd59-2937d90c1f30 req-ad5bb27a-91db-40e4-a890-2fbdcebba064 service nova] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Updating instance_info_cache with network_info: [{"id": "2df835dc-fa26-45ab-9d30-58f85c9e3331", "address": "fa:16:3e:d9:6c:e7", "network": {"id": "0c0dd1a7-6687-45c5-82b1-33f34bdbb848", "bridge": "br-int", "label": "tempest-ServersTestJSON-523240467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c8959d2f1fe4706b7407c11e128b773", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffc811e4-0e4e-4748-8c7e-b3f14ccbd42d", "external-id": "nsx-vlan-transportzone-404", "segmentation_id": 404, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2df835dc-fa", "ovs_interfaceid": "2df835dc-fa26-45ab-9d30-58f85c9e3331", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.841296] env[63021]: DEBUG oslo_vmware.api [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294153, 'name': ReconfigVM_Task, 'duration_secs': 0.275691} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.841590] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Reconfigured VM instance instance-0000005e to attach disk [datastore2] 67aed8eb-d08c-47cc-8558-5c5a0f058d6e/67aed8eb-d08c-47cc-8558-5c5a0f058d6e.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1006.842264] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-941e8b3a-5737-44a2-9546-4f9d02858ddc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.849488] env[63021]: DEBUG oslo_vmware.api [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1006.849488] env[63021]: value = "task-1294154" [ 1006.849488] env[63021]: _type = "Task" [ 1006.849488] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.858280] env[63021]: DEBUG oslo_vmware.api [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294154, 'name': Rename_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.981101] env[63021]: DEBUG nova.scheduler.client.report [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1007.224872] env[63021]: WARNING nova.network.neutron [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] 2c4f9111-2f39-4bd0-95bc-09b96d535ea3 already exists in list: networks containing: ['2c4f9111-2f39-4bd0-95bc-09b96d535ea3']. ignoring it [ 1007.234724] env[63021]: DEBUG nova.network.neutron [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1007.277520] env[63021]: DEBUG oslo_vmware.api [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]529ec01f-324d-8559-3382-57bd2badf8c3, 'name': SearchDatastore_Task, 'duration_secs': 0.009558} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.278033] env[63021]: DEBUG oslo_concurrency.lockutils [req-2a271628-3230-421f-bd59-2937d90c1f30 req-ad5bb27a-91db-40e4-a890-2fbdcebba064 service nova] Releasing lock "refresh_cache-023c0891-2b08-48bb-9754-408ce7d04c9c" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.278399] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.278647] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 023c0891-2b08-48bb-9754-408ce7d04c9c/023c0891-2b08-48bb-9754-408ce7d04c9c.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1007.278920] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d884d49a-63d7-420e-85d1-937a26a07c18 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.288119] env[63021]: DEBUG oslo_vmware.api [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1007.288119] env[63021]: value = "task-1294155" [ 1007.288119] env[63021]: _type = "Task" [ 1007.288119] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.297946] env[63021]: DEBUG oslo_vmware.api [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294155, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.360609] env[63021]: DEBUG oslo_vmware.api [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294154, 'name': Rename_Task, 'duration_secs': 0.145886} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.360792] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1007.361122] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b42813e9-6d73-40c4-8843-c00a1a0e0779 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.368676] env[63021]: DEBUG oslo_vmware.api [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1007.368676] env[63021]: value = "task-1294156" [ 1007.368676] env[63021]: _type = "Task" [ 1007.368676] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.374274] env[63021]: DEBUG nova.network.neutron [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Updating instance_info_cache with network_info: [{"id": "4a29562f-684a-4738-98b0-f1281b0bc4c3", "address": "fa:16:3e:48:75:02", "network": {"id": "81917a55-4e41-4898-9f4e-4fced52b5a86", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1793837736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e238d5c1668540d2a9fffd3fd832d9b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a29562f-68", "ovs_interfaceid": "4a29562f-684a-4738-98b0-f1281b0bc4c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.379144] env[63021]: DEBUG oslo_vmware.api [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294156, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.486446] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.887s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.510487] env[63021]: INFO nova.scheduler.client.report [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Deleted allocations for instance 9f822151-3307-4093-a972-8af128ea1892 [ 1007.799596] env[63021]: DEBUG oslo_vmware.api [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294155, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.867147] env[63021]: DEBUG nova.network.neutron [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Updating instance_info_cache with network_info: [{"id": "3752044a-9145-469d-8edb-6255d7d0f928", "address": "fa:16:3e:fe:57:1f", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3752044a-91", "ovs_interfaceid": "3752044a-9145-469d-8edb-6255d7d0f928", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "01e26b49-58a2-411d-b025-b1ca941026c4", "address": "fa:16:3e:1d:da:30", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01e26b49-58", "ovs_interfaceid": "01e26b49-58a2-411d-b025-b1ca941026c4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.878432] env[63021]: DEBUG oslo_vmware.api [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294156, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.880393] env[63021]: DEBUG oslo_concurrency.lockutils [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Releasing lock "refresh_cache-07510760-13ff-4797-86ca-86b384aacdfa" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.880853] env[63021]: DEBUG nova.compute.manager [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Instance network_info: |[{"id": "4a29562f-684a-4738-98b0-f1281b0bc4c3", "address": "fa:16:3e:48:75:02", "network": {"id": "81917a55-4e41-4898-9f4e-4fced52b5a86", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1793837736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e238d5c1668540d2a9fffd3fd832d9b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a29562f-68", "ovs_interfaceid": "4a29562f-684a-4738-98b0-f1281b0bc4c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1007.881911] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:48:75:02', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '535b175f-71d3-4226-81ab-ca253f27fedd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4a29562f-684a-4738-98b0-f1281b0bc4c3', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1007.889698] env[63021]: DEBUG oslo.service.loopingcall [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1007.890159] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1007.890385] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9b9deb69-257a-40ed-8117-ec5b38b2e95e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.910189] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1007.910189] env[63021]: value = "task-1294157" [ 1007.910189] env[63021]: _type = "Task" [ 1007.910189] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.918034] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294157, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.020323] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2be46a9a-298e-4c9e-aaa9-b413e6e05f8c tempest-ServerDiskConfigTestJSON-327740150 tempest-ServerDiskConfigTestJSON-327740150-project-member] Lock "9f822151-3307-4093-a972-8af128ea1892" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.111s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.299713] env[63021]: DEBUG oslo_vmware.api [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294155, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.513934} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.300393] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 023c0891-2b08-48bb-9754-408ce7d04c9c/023c0891-2b08-48bb-9754-408ce7d04c9c.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1008.300612] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1008.300873] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8855aad5-50a0-4f13-96ed-2f926626543d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.306733] env[63021]: DEBUG oslo_vmware.api [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1008.306733] env[63021]: value = "task-1294158" [ 1008.306733] env[63021]: _type = "Task" [ 1008.306733] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.314851] env[63021]: DEBUG oslo_vmware.api [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294158, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.362867] env[63021]: DEBUG nova.compute.manager [req-03af9705-9963-4204-8683-87131d076480 req-9c3dba30-b56d-4285-92f4-1d65659ed26f service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Received event network-vif-plugged-01e26b49-58a2-411d-b025-b1ca941026c4 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1008.363115] env[63021]: DEBUG oslo_concurrency.lockutils [req-03af9705-9963-4204-8683-87131d076480 req-9c3dba30-b56d-4285-92f4-1d65659ed26f service nova] Acquiring lock "6155c098-e4a6-47e6-b343-4a77ca90eb2e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.363370] env[63021]: DEBUG oslo_concurrency.lockutils [req-03af9705-9963-4204-8683-87131d076480 req-9c3dba30-b56d-4285-92f4-1d65659ed26f service nova] Lock "6155c098-e4a6-47e6-b343-4a77ca90eb2e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.363606] env[63021]: DEBUG oslo_concurrency.lockutils [req-03af9705-9963-4204-8683-87131d076480 req-9c3dba30-b56d-4285-92f4-1d65659ed26f service nova] Lock "6155c098-e4a6-47e6-b343-4a77ca90eb2e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.363794] env[63021]: DEBUG nova.compute.manager [req-03af9705-9963-4204-8683-87131d076480 req-9c3dba30-b56d-4285-92f4-1d65659ed26f service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] No waiting events found dispatching network-vif-plugged-01e26b49-58a2-411d-b025-b1ca941026c4 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1008.363954] env[63021]: WARNING nova.compute.manager [req-03af9705-9963-4204-8683-87131d076480 req-9c3dba30-b56d-4285-92f4-1d65659ed26f service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Received unexpected event network-vif-plugged-01e26b49-58a2-411d-b025-b1ca941026c4 for instance with vm_state active and task_state None. [ 1008.364128] env[63021]: DEBUG nova.compute.manager [req-03af9705-9963-4204-8683-87131d076480 req-9c3dba30-b56d-4285-92f4-1d65659ed26f service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Received event network-changed-01e26b49-58a2-411d-b025-b1ca941026c4 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1008.364291] env[63021]: DEBUG nova.compute.manager [req-03af9705-9963-4204-8683-87131d076480 req-9c3dba30-b56d-4285-92f4-1d65659ed26f service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Refreshing instance network info cache due to event network-changed-01e26b49-58a2-411d-b025-b1ca941026c4. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1008.364451] env[63021]: DEBUG oslo_concurrency.lockutils [req-03af9705-9963-4204-8683-87131d076480 req-9c3dba30-b56d-4285-92f4-1d65659ed26f service nova] Acquiring lock "refresh_cache-6155c098-e4a6-47e6-b343-4a77ca90eb2e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.369658] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Releasing lock "refresh_cache-6155c098-e4a6-47e6-b343-4a77ca90eb2e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.370327] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "6155c098-e4a6-47e6-b343-4a77ca90eb2e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.370518] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquired lock "6155c098-e4a6-47e6-b343-4a77ca90eb2e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.370843] env[63021]: DEBUG oslo_concurrency.lockutils [req-03af9705-9963-4204-8683-87131d076480 req-9c3dba30-b56d-4285-92f4-1d65659ed26f service nova] Acquired lock "refresh_cache-6155c098-e4a6-47e6-b343-4a77ca90eb2e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.371030] env[63021]: DEBUG nova.network.neutron [req-03af9705-9963-4204-8683-87131d076480 req-9c3dba30-b56d-4285-92f4-1d65659ed26f service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Refreshing network info cache for port 01e26b49-58a2-411d-b025-b1ca941026c4 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1008.372667] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0d1384f-4ae6-4300-b25c-692f9d5ac0bd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.385090] env[63021]: DEBUG oslo_vmware.api [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294156, 'name': PowerOnVM_Task, 'duration_secs': 0.724902} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.396572] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1008.396572] env[63021]: INFO nova.compute.manager [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Took 7.09 seconds to spawn the instance on the hypervisor. [ 1008.396572] env[63021]: DEBUG nova.compute.manager [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1008.398186] env[63021]: DEBUG nova.virt.hardware [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1008.398410] env[63021]: DEBUG nova.virt.hardware [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1008.398567] env[63021]: DEBUG nova.virt.hardware [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1008.398748] env[63021]: DEBUG nova.virt.hardware [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1008.398894] env[63021]: DEBUG nova.virt.hardware [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1008.399082] env[63021]: DEBUG nova.virt.hardware [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1008.399300] env[63021]: DEBUG nova.virt.hardware [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1008.399457] env[63021]: DEBUG nova.virt.hardware [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1008.399621] env[63021]: DEBUG nova.virt.hardware [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1008.399786] env[63021]: DEBUG nova.virt.hardware [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1008.399954] env[63021]: DEBUG nova.virt.hardware [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1008.406149] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Reconfiguring VM to attach interface {{(pid=63021) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1008.406930] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-564e78ef-8d34-400c-8259-0e31929411a1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.410236] env[63021]: DEBUG nova.compute.manager [req-7a3e65eb-c973-4d63-81a3-44618d85c2c6 req-06c0786f-2d28-43d7-8399-b39b3c0e19b3 service nova] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Received event network-changed-4a29562f-684a-4738-98b0-f1281b0bc4c3 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1008.410413] env[63021]: DEBUG nova.compute.manager [req-7a3e65eb-c973-4d63-81a3-44618d85c2c6 req-06c0786f-2d28-43d7-8399-b39b3c0e19b3 service nova] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Refreshing instance network info cache due to event network-changed-4a29562f-684a-4738-98b0-f1281b0bc4c3. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1008.410623] env[63021]: DEBUG oslo_concurrency.lockutils [req-7a3e65eb-c973-4d63-81a3-44618d85c2c6 req-06c0786f-2d28-43d7-8399-b39b3c0e19b3 service nova] Acquiring lock "refresh_cache-07510760-13ff-4797-86ca-86b384aacdfa" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.410767] env[63021]: DEBUG oslo_concurrency.lockutils [req-7a3e65eb-c973-4d63-81a3-44618d85c2c6 req-06c0786f-2d28-43d7-8399-b39b3c0e19b3 service nova] Acquired lock "refresh_cache-07510760-13ff-4797-86ca-86b384aacdfa" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.410923] env[63021]: DEBUG nova.network.neutron [req-7a3e65eb-c973-4d63-81a3-44618d85c2c6 req-06c0786f-2d28-43d7-8399-b39b3c0e19b3 service nova] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Refreshing network info cache for port 4a29562f-684a-4738-98b0-f1281b0bc4c3 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1008.412562] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2d825c72-0f87-49b0-9941-df909bc0e6e4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.436519] env[63021]: DEBUG oslo_vmware.api [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 1008.436519] env[63021]: value = "task-1294159" [ 1008.436519] env[63021]: _type = "Task" [ 1008.436519] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.442950] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294157, 'name': CreateVM_Task, 'duration_secs': 0.415892} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.443412] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1008.444071] env[63021]: DEBUG oslo_concurrency.lockutils [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.444238] env[63021]: DEBUG oslo_concurrency.lockutils [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.444548] env[63021]: DEBUG oslo_concurrency.lockutils [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1008.445235] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c74d6cf7-1191-496f-bcb8-f415e578a1f9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.449463] env[63021]: DEBUG oslo_vmware.api [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294159, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.452445] env[63021]: DEBUG oslo_vmware.api [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1008.452445] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]520e1da1-bb6a-4a31-5087-185fd1578748" [ 1008.452445] env[63021]: _type = "Task" [ 1008.452445] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.460804] env[63021]: DEBUG oslo_vmware.api [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]520e1da1-bb6a-4a31-5087-185fd1578748, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.735914] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Acquiring lock "45c9e2be-eb19-4b83-b280-c9eeaddfccba" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.736750] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lock "45c9e2be-eb19-4b83-b280-c9eeaddfccba" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.736994] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Acquiring lock "45c9e2be-eb19-4b83-b280-c9eeaddfccba-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.737213] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lock "45c9e2be-eb19-4b83-b280-c9eeaddfccba-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.737394] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lock "45c9e2be-eb19-4b83-b280-c9eeaddfccba-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.739667] env[63021]: INFO nova.compute.manager [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Terminating instance [ 1008.741861] env[63021]: DEBUG nova.compute.manager [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1008.742471] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1008.743818] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f249b5e8-7bba-4a61-947d-a0b32a9e30b8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.752447] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1008.753803] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-767482bd-0fa6-4fe6-b729-89901c717194 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.758442] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquiring lock "8d9387b8-6a84-4ada-8ac7-4ffd922b675a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.758750] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "8d9387b8-6a84-4ada-8ac7-4ffd922b675a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.764763] env[63021]: DEBUG oslo_vmware.api [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Waiting for the task: (returnval){ [ 1008.764763] env[63021]: value = "task-1294160" [ 1008.764763] env[63021]: _type = "Task" [ 1008.764763] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.773769] env[63021]: DEBUG oslo_vmware.api [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1294160, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.817225] env[63021]: DEBUG oslo_vmware.api [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294158, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.186012} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.817500] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1008.818286] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b626e0f-50ad-43ab-9b03-3b0cb5fa46d3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.840301] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] 023c0891-2b08-48bb-9754-408ce7d04c9c/023c0891-2b08-48bb-9754-408ce7d04c9c.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1008.840809] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c23f399-8e4a-4719-9629-33a2037f2546 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.861457] env[63021]: DEBUG oslo_vmware.api [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1008.861457] env[63021]: value = "task-1294161" [ 1008.861457] env[63021]: _type = "Task" [ 1008.861457] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.869676] env[63021]: DEBUG oslo_vmware.api [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294161, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.951200] env[63021]: INFO nova.compute.manager [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Took 11.95 seconds to build instance. [ 1008.961832] env[63021]: DEBUG oslo_vmware.api [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294159, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.971534] env[63021]: DEBUG oslo_vmware.api [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]520e1da1-bb6a-4a31-5087-185fd1578748, 'name': SearchDatastore_Task, 'duration_secs': 0.019384} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.971901] env[63021]: DEBUG oslo_concurrency.lockutils [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.972230] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1008.972563] env[63021]: DEBUG oslo_concurrency.lockutils [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.972799] env[63021]: DEBUG oslo_concurrency.lockutils [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.973080] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1008.973418] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0bdde922-cbf4-4540-9aa6-39b623f6c2c2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.988696] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1008.988943] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1008.989923] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2883a9fe-f0b7-44ef-af63-0e7c03addf12 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.996346] env[63021]: DEBUG oslo_vmware.api [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1008.996346] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52b6c40f-270d-8a90-83c4-4e29a62908a8" [ 1008.996346] env[63021]: _type = "Task" [ 1008.996346] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.007762] env[63021]: DEBUG oslo_vmware.api [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52b6c40f-270d-8a90-83c4-4e29a62908a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.166547] env[63021]: DEBUG nova.network.neutron [req-7a3e65eb-c973-4d63-81a3-44618d85c2c6 req-06c0786f-2d28-43d7-8399-b39b3c0e19b3 service nova] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Updated VIF entry in instance network info cache for port 4a29562f-684a-4738-98b0-f1281b0bc4c3. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1009.167142] env[63021]: DEBUG nova.network.neutron [req-7a3e65eb-c973-4d63-81a3-44618d85c2c6 req-06c0786f-2d28-43d7-8399-b39b3c0e19b3 service nova] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Updating instance_info_cache with network_info: [{"id": "4a29562f-684a-4738-98b0-f1281b0bc4c3", "address": "fa:16:3e:48:75:02", "network": {"id": "81917a55-4e41-4898-9f4e-4fced52b5a86", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1793837736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e238d5c1668540d2a9fffd3fd832d9b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a29562f-68", "ovs_interfaceid": "4a29562f-684a-4738-98b0-f1281b0bc4c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.214331] env[63021]: DEBUG nova.network.neutron [req-03af9705-9963-4204-8683-87131d076480 req-9c3dba30-b56d-4285-92f4-1d65659ed26f service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Updated VIF entry in instance network info cache for port 01e26b49-58a2-411d-b025-b1ca941026c4. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1009.214868] env[63021]: DEBUG nova.network.neutron [req-03af9705-9963-4204-8683-87131d076480 req-9c3dba30-b56d-4285-92f4-1d65659ed26f service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Updating instance_info_cache with network_info: [{"id": "3752044a-9145-469d-8edb-6255d7d0f928", "address": "fa:16:3e:fe:57:1f", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3752044a-91", "ovs_interfaceid": "3752044a-9145-469d-8edb-6255d7d0f928", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "01e26b49-58a2-411d-b025-b1ca941026c4", "address": "fa:16:3e:1d:da:30", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01e26b49-58", "ovs_interfaceid": "01e26b49-58a2-411d-b025-b1ca941026c4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.264052] env[63021]: DEBUG nova.compute.manager [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1009.276554] env[63021]: DEBUG oslo_vmware.api [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1294160, 'name': PowerOffVM_Task, 'duration_secs': 0.359768} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.276873] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1009.277077] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1009.277344] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cb5c83f0-6733-4568-aa37-598ad81a67e0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.352690] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1009.353170] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Deleting contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1009.353170] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Deleting the datastore file [datastore2] 45c9e2be-eb19-4b83-b280-c9eeaddfccba {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1009.353387] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-75e51637-9b89-4b67-86de-61f8bb99bd9d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.359876] env[63021]: DEBUG oslo_vmware.api [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Waiting for the task: (returnval){ [ 1009.359876] env[63021]: value = "task-1294163" [ 1009.359876] env[63021]: _type = "Task" [ 1009.359876] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.370495] env[63021]: DEBUG oslo_vmware.api [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1294163, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.374062] env[63021]: DEBUG oslo_vmware.api [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294161, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.452360] env[63021]: DEBUG oslo_vmware.api [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294159, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.453902] env[63021]: DEBUG oslo_concurrency.lockutils [None req-cc5d3d7b-74a9-49db-b9b3-547fdd90c192 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "67aed8eb-d08c-47cc-8558-5c5a0f058d6e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.460s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.507024] env[63021]: DEBUG oslo_vmware.api [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52b6c40f-270d-8a90-83c4-4e29a62908a8, 'name': SearchDatastore_Task, 'duration_secs': 0.045158} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.507836] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a40d79f-805e-4a6f-bc45-07bec5a83283 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.512891] env[63021]: DEBUG oslo_vmware.api [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1009.512891] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52dddc38-b1b8-5d2c-beb7-42b68a08f9d5" [ 1009.512891] env[63021]: _type = "Task" [ 1009.512891] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.520344] env[63021]: DEBUG oslo_vmware.api [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52dddc38-b1b8-5d2c-beb7-42b68a08f9d5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.669906] env[63021]: DEBUG oslo_concurrency.lockutils [req-7a3e65eb-c973-4d63-81a3-44618d85c2c6 req-06c0786f-2d28-43d7-8399-b39b3c0e19b3 service nova] Releasing lock "refresh_cache-07510760-13ff-4797-86ca-86b384aacdfa" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.718048] env[63021]: DEBUG oslo_concurrency.lockutils [req-03af9705-9963-4204-8683-87131d076480 req-9c3dba30-b56d-4285-92f4-1d65659ed26f service nova] Releasing lock "refresh_cache-6155c098-e4a6-47e6-b343-4a77ca90eb2e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.799194] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.799466] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.801093] env[63021]: INFO nova.compute.claims [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1009.879506] env[63021]: DEBUG oslo_vmware.api [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294161, 'name': ReconfigVM_Task, 'duration_secs': 0.637655} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.881119] env[63021]: DEBUG oslo_vmware.api [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Task: {'id': task-1294163, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.221716} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.881419] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Reconfigured VM instance instance-0000005f to attach disk [datastore2] 023c0891-2b08-48bb-9754-408ce7d04c9c/023c0891-2b08-48bb-9754-408ce7d04c9c.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1009.882070] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1009.882262] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Deleted contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1009.882437] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1009.882611] env[63021]: INFO nova.compute.manager [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1009.882845] env[63021]: DEBUG oslo.service.loopingcall [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1009.883062] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b361532a-1cfc-4571-a3ec-3beea9ccb274 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.886884] env[63021]: DEBUG nova.compute.manager [-] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1009.886978] env[63021]: DEBUG nova.network.neutron [-] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1009.896551] env[63021]: DEBUG oslo_vmware.api [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1009.896551] env[63021]: value = "task-1294164" [ 1009.896551] env[63021]: _type = "Task" [ 1009.896551] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.905612] env[63021]: DEBUG oslo_vmware.api [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294164, 'name': Rename_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.952020] env[63021]: DEBUG oslo_vmware.api [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294159, 'name': ReconfigVM_Task, 'duration_secs': 1.169615} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.952777] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Releasing lock "6155c098-e4a6-47e6-b343-4a77ca90eb2e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.953141] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Reconfigured VM to attach interface {{(pid=63021) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1010.024924] env[63021]: DEBUG oslo_vmware.api [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52dddc38-b1b8-5d2c-beb7-42b68a08f9d5, 'name': SearchDatastore_Task, 'duration_secs': 0.008718} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.025445] env[63021]: DEBUG oslo_concurrency.lockutils [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.025770] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 07510760-13ff-4797-86ca-86b384aacdfa/07510760-13ff-4797-86ca-86b384aacdfa.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1010.026084] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bd16846e-90f4-43d6-83ae-745f3de3766b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.033112] env[63021]: DEBUG oslo_vmware.api [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1010.033112] env[63021]: value = "task-1294165" [ 1010.033112] env[63021]: _type = "Task" [ 1010.033112] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.042024] env[63021]: DEBUG oslo_vmware.api [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294165, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.408550] env[63021]: DEBUG oslo_vmware.api [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294164, 'name': Rename_Task, 'duration_secs': 0.259522} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.408898] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1010.409195] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fa0720ae-1a88-4515-b8c3-db75d0c7fbe0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.418249] env[63021]: DEBUG oslo_vmware.api [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1010.418249] env[63021]: value = "task-1294166" [ 1010.418249] env[63021]: _type = "Task" [ 1010.418249] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.426873] env[63021]: DEBUG oslo_vmware.api [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294166, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.458061] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4f1f5ed6-71c1-4a1d-8df5-7d0ad706f1af tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "interface-6155c098-e4a6-47e6-b343-4a77ca90eb2e-01e26b49-58a2-411d-b025-b1ca941026c4" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.244s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.550226] env[63021]: DEBUG oslo_vmware.api [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294165, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.683846] env[63021]: DEBUG nova.compute.manager [req-cfac1d63-c882-40ed-85b2-df68b37e102b req-607c0a93-16a5-4b11-ac51-11a2787d0a69 service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Received event network-vif-deleted-43e48c53-d06d-41ff-97dd-a7747d5551d9 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1010.683846] env[63021]: INFO nova.compute.manager [req-cfac1d63-c882-40ed-85b2-df68b37e102b req-607c0a93-16a5-4b11-ac51-11a2787d0a69 service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Neutron deleted interface 43e48c53-d06d-41ff-97dd-a7747d5551d9; detaching it from the instance and deleting it from the info cache [ 1010.683846] env[63021]: DEBUG nova.network.neutron [req-cfac1d63-c882-40ed-85b2-df68b37e102b req-607c0a93-16a5-4b11-ac51-11a2787d0a69 service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.708255] env[63021]: INFO nova.compute.manager [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Rebuilding instance [ 1010.761755] env[63021]: DEBUG nova.compute.manager [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1010.762755] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74a1c8d1-da4b-4b82-a37e-bc5e6dbd5ec5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.936959] env[63021]: DEBUG oslo_vmware.api [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294166, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.016906] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21bfcb22-6ba1-4f04-ad99-25bfbe9d8c56 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.024786] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0e6606c-161f-4976-a88c-beb8809265de {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.064256] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aae0b26-f4e8-4ef6-be21-2d58bd27b0fc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.075987] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f11a80bc-8df4-4026-be19-ba98526393fe {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.081151] env[63021]: DEBUG oslo_vmware.api [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294165, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.730131} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.082093] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 07510760-13ff-4797-86ca-86b384aacdfa/07510760-13ff-4797-86ca-86b384aacdfa.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1011.082526] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1011.083127] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6e0eb92c-125d-4987-b5c2-4048a11c15cf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.095078] env[63021]: DEBUG nova.compute.provider_tree [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1011.102373] env[63021]: DEBUG oslo_vmware.api [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1011.102373] env[63021]: value = "task-1294167" [ 1011.102373] env[63021]: _type = "Task" [ 1011.102373] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.114035] env[63021]: DEBUG oslo_vmware.api [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294167, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.158451] env[63021]: DEBUG nova.network.neutron [-] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.186878] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3cb28f63-b9a3-4a78-befe-fdbece877bb5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.196099] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94a6dac6-724f-4319-81bd-e8698ffa20f6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.226626] env[63021]: DEBUG nova.compute.manager [req-cfac1d63-c882-40ed-85b2-df68b37e102b req-607c0a93-16a5-4b11-ac51-11a2787d0a69 service nova] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Detach interface failed, port_id=43e48c53-d06d-41ff-97dd-a7747d5551d9, reason: Instance 45c9e2be-eb19-4b83-b280-c9eeaddfccba could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1011.276924] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1011.277863] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b4676c28-706a-4330-be7c-b1587deffb25 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.285977] env[63021]: DEBUG oslo_vmware.api [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1011.285977] env[63021]: value = "task-1294168" [ 1011.285977] env[63021]: _type = "Task" [ 1011.285977] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.298774] env[63021]: DEBUG oslo_vmware.api [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294168, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.429453] env[63021]: DEBUG oslo_vmware.api [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294166, 'name': PowerOnVM_Task, 'duration_secs': 0.762965} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.429835] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1011.429835] env[63021]: INFO nova.compute.manager [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Took 7.83 seconds to spawn the instance on the hypervisor. [ 1011.430480] env[63021]: DEBUG nova.compute.manager [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1011.430748] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f21155b7-3d82-4f55-b095-cbcc70e8f03d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.598990] env[63021]: DEBUG nova.scheduler.client.report [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1011.613407] env[63021]: DEBUG oslo_vmware.api [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294167, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.090628} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.613555] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1011.614331] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1b5404c-3b37-4b95-bdb3-774c650ca31d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.637602] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] 07510760-13ff-4797-86ca-86b384aacdfa/07510760-13ff-4797-86ca-86b384aacdfa.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1011.638155] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2980c64b-f7bc-4612-aa88-7a4078a40572 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.657568] env[63021]: DEBUG oslo_vmware.api [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1011.657568] env[63021]: value = "task-1294169" [ 1011.657568] env[63021]: _type = "Task" [ 1011.657568] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.661014] env[63021]: INFO nova.compute.manager [-] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Took 1.77 seconds to deallocate network for instance. [ 1011.668761] env[63021]: DEBUG oslo_vmware.api [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294169, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.798722] env[63021]: DEBUG oslo_vmware.api [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294168, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.952674] env[63021]: DEBUG oslo_concurrency.lockutils [None req-8645ea1e-dd6a-47bb-be9d-3069b9df0403 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "interface-6155c098-e4a6-47e6-b343-4a77ca90eb2e-01e26b49-58a2-411d-b025-b1ca941026c4" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.952983] env[63021]: DEBUG oslo_concurrency.lockutils [None req-8645ea1e-dd6a-47bb-be9d-3069b9df0403 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "interface-6155c098-e4a6-47e6-b343-4a77ca90eb2e-01e26b49-58a2-411d-b025-b1ca941026c4" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.954482] env[63021]: INFO nova.compute.manager [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Took 13.39 seconds to build instance. [ 1012.108997] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.309s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.109556] env[63021]: DEBUG nova.compute.manager [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1012.169120] env[63021]: DEBUG oslo_vmware.api [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294169, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.170091] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.170338] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.170551] env[63021]: DEBUG nova.objects.instance [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lazy-loading 'resources' on Instance uuid 45c9e2be-eb19-4b83-b280-c9eeaddfccba {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1012.298044] env[63021]: DEBUG oslo_vmware.api [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294168, 'name': PowerOffVM_Task, 'duration_secs': 0.566804} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.298396] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1012.298642] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1012.299461] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-692bc237-6d52-491d-b3e3-9a44d67a2183 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.306144] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1012.306769] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-35fa1c9a-4643-4a4d-a4e4-3365bcf32b60 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.373094] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1012.373343] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Deleting contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1012.373527] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Deleting the datastore file [datastore2] 67aed8eb-d08c-47cc-8558-5c5a0f058d6e {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1012.373803] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9d8a0a73-d06a-4171-901a-7c3061fcd8cf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.380382] env[63021]: DEBUG oslo_vmware.api [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1012.380382] env[63021]: value = "task-1294171" [ 1012.380382] env[63021]: _type = "Task" [ 1012.380382] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.388309] env[63021]: DEBUG oslo_vmware.api [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294171, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.456984] env[63021]: DEBUG oslo_concurrency.lockutils [None req-8645ea1e-dd6a-47bb-be9d-3069b9df0403 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "6155c098-e4a6-47e6-b343-4a77ca90eb2e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.457500] env[63021]: DEBUG oslo_concurrency.lockutils [None req-8645ea1e-dd6a-47bb-be9d-3069b9df0403 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquired lock "6155c098-e4a6-47e6-b343-4a77ca90eb2e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.457713] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0cbd383-68bd-454f-ad55-cf26586239de tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "023c0891-2b08-48bb-9754-408ce7d04c9c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.902s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.458505] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-523e2e89-07e3-4b5b-87af-7fc009b24d41 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.476660] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8cfd1bd-6d06-46b4-8b82-08d153f8459f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.502749] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-8645ea1e-dd6a-47bb-be9d-3069b9df0403 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Reconfiguring VM to detach interface {{(pid=63021) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1012.503057] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1079b5cc-6881-401a-bd75-90e0c8555dcd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.520615] env[63021]: DEBUG oslo_vmware.api [None req-8645ea1e-dd6a-47bb-be9d-3069b9df0403 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 1012.520615] env[63021]: value = "task-1294172" [ 1012.520615] env[63021]: _type = "Task" [ 1012.520615] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.528297] env[63021]: DEBUG oslo_vmware.api [None req-8645ea1e-dd6a-47bb-be9d-3069b9df0403 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294172, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.614859] env[63021]: DEBUG nova.compute.utils [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1012.616545] env[63021]: DEBUG nova.compute.manager [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1012.616745] env[63021]: DEBUG nova.network.neutron [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1012.667627] env[63021]: DEBUG oslo_vmware.api [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294169, 'name': ReconfigVM_Task, 'duration_secs': 0.860236} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.667949] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Reconfigured VM instance instance-00000060 to attach disk [datastore2] 07510760-13ff-4797-86ca-86b384aacdfa/07510760-13ff-4797-86ca-86b384aacdfa.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1012.669616] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a8edc07c-3aa2-41b5-b9d3-e30094a3a99b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.675984] env[63021]: DEBUG oslo_vmware.api [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1012.675984] env[63021]: value = "task-1294173" [ 1012.675984] env[63021]: _type = "Task" [ 1012.675984] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.685344] env[63021]: DEBUG oslo_vmware.api [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294173, 'name': Rename_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.696319] env[63021]: DEBUG nova.policy [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c8c1392321694d17bc08633dfbfe5ee2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5661c44b534b4a7e9452e914fc80d66e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 1012.842620] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5001de9a-02be-44c8-84a6-fb79c3fd88e7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.853101] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e72b3fcf-acb0-4a86-87a5-c476e07be04e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.637908] env[63021]: DEBUG nova.compute.manager [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1013.642925] env[63021]: DEBUG nova.network.neutron [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Successfully created port: 1bb789b9-f77e-4e8c-a00b-c513af834d07 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1013.659074] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70d519d2-d3c0-4a83-aa40-7dcde4a6ba85 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.664674] env[63021]: DEBUG oslo_vmware.api [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294171, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.330264} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.672349] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1013.672554] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Deleted contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1013.672729] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1013.675593] env[63021]: DEBUG oslo_vmware.api [None req-8645ea1e-dd6a-47bb-be9d-3069b9df0403 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294172, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.676326] env[63021]: DEBUG oslo_vmware.api [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294173, 'name': Rename_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.677221] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b72acbb0-8931-4bae-bf12-82d3df25bd67 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.692389] env[63021]: DEBUG nova.compute.provider_tree [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1014.160566] env[63021]: DEBUG oslo_concurrency.lockutils [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "023c0891-2b08-48bb-9754-408ce7d04c9c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.161010] env[63021]: DEBUG oslo_concurrency.lockutils [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "023c0891-2b08-48bb-9754-408ce7d04c9c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.161334] env[63021]: DEBUG oslo_concurrency.lockutils [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "023c0891-2b08-48bb-9754-408ce7d04c9c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.161628] env[63021]: DEBUG oslo_concurrency.lockutils [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "023c0891-2b08-48bb-9754-408ce7d04c9c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.161897] env[63021]: DEBUG oslo_concurrency.lockutils [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "023c0891-2b08-48bb-9754-408ce7d04c9c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.170682] env[63021]: DEBUG oslo_vmware.api [None req-8645ea1e-dd6a-47bb-be9d-3069b9df0403 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294172, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.171370] env[63021]: INFO nova.compute.manager [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Terminating instance [ 1014.175873] env[63021]: DEBUG oslo_vmware.api [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294173, 'name': Rename_Task, 'duration_secs': 1.144213} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.176144] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1014.179397] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3e045663-5cf6-441b-b860-41ebc46ecff8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.181412] env[63021]: DEBUG nova.compute.manager [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1014.181608] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1014.182517] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d964a7f-8d43-4882-9004-ef9f3d2bacce {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.189745] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1014.190889] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1260e3ea-d18b-4593-915f-0eb961f9c141 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.192440] env[63021]: DEBUG oslo_vmware.api [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1014.192440] env[63021]: value = "task-1294174" [ 1014.192440] env[63021]: _type = "Task" [ 1014.192440] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.198059] env[63021]: DEBUG nova.scheduler.client.report [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1014.201500] env[63021]: DEBUG oslo_vmware.api [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1014.201500] env[63021]: value = "task-1294175" [ 1014.201500] env[63021]: _type = "Task" [ 1014.201500] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.211489] env[63021]: DEBUG oslo_vmware.api [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294174, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.217825] env[63021]: DEBUG oslo_vmware.api [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294175, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.649557] env[63021]: DEBUG oslo_vmware.api [None req-8645ea1e-dd6a-47bb-be9d-3069b9df0403 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294172, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.661893] env[63021]: DEBUG nova.compute.manager [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1014.695393] env[63021]: DEBUG nova.virt.hardware [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1014.695393] env[63021]: DEBUG nova.virt.hardware [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1014.695583] env[63021]: DEBUG nova.virt.hardware [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1014.695667] env[63021]: DEBUG nova.virt.hardware [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1014.695841] env[63021]: DEBUG nova.virt.hardware [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1014.696501] env[63021]: DEBUG nova.virt.hardware [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1014.696820] env[63021]: DEBUG nova.virt.hardware [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1014.697014] env[63021]: DEBUG nova.virt.hardware [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1014.697204] env[63021]: DEBUG nova.virt.hardware [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1014.697371] env[63021]: DEBUG nova.virt.hardware [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1014.697545] env[63021]: DEBUG nova.virt.hardware [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1014.698501] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edea078b-88cb-4acb-a5ff-a54e7c1f46a5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.704712] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.534s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.721486] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3daee37-e637-4c33-b456-a7dedf0aabc3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.726376] env[63021]: DEBUG oslo_vmware.api [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294174, 'name': PowerOnVM_Task, 'duration_secs': 0.447407} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.730494] env[63021]: DEBUG nova.virt.hardware [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1014.730642] env[63021]: DEBUG nova.virt.hardware [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1014.731016] env[63021]: DEBUG nova.virt.hardware [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1014.731597] env[63021]: DEBUG nova.virt.hardware [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1014.731831] env[63021]: DEBUG nova.virt.hardware [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1014.732180] env[63021]: DEBUG nova.virt.hardware [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1014.732597] env[63021]: DEBUG nova.virt.hardware [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1014.732938] env[63021]: DEBUG nova.virt.hardware [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1014.733259] env[63021]: DEBUG nova.virt.hardware [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1014.733635] env[63021]: DEBUG nova.virt.hardware [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1014.734550] env[63021]: DEBUG nova.virt.hardware [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1014.738705] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1014.738705] env[63021]: INFO nova.compute.manager [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Took 8.86 seconds to spawn the instance on the hypervisor. [ 1014.738705] env[63021]: DEBUG nova.compute.manager [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1014.738705] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c5b3433-20c5-4c04-b51f-7f7e129216d5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.742182] env[63021]: DEBUG oslo_vmware.api [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294175, 'name': PowerOffVM_Task, 'duration_secs': 0.165738} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.742560] env[63021]: INFO nova.scheduler.client.report [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Deleted allocations for instance 45c9e2be-eb19-4b83-b280-c9eeaddfccba [ 1014.744420] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9bf9d8e-4391-44d8-b136-d72cc3cdefd0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.752660] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1014.752660] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1014.762051] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a94e6b68-5796-4e8c-8be3-a52c7cb1607c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.770021] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d935aa89-179d-4e43-9eee-675f259a5c37 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.787841] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6b:ae:03', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '371c655a-b09c-48dc-ad0a-da708503e86e', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1014.796748] env[63021]: DEBUG oslo.service.loopingcall [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1014.797091] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1014.797328] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-28da4224-12fa-42cc-a3c2-be2c32f36f53 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.818935] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1014.818935] env[63021]: value = "task-1294177" [ 1014.818935] env[63021]: _type = "Task" [ 1014.818935] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.825946] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1014.825946] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Deleting contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1014.825946] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Deleting the datastore file [datastore2] 023c0891-2b08-48bb-9754-408ce7d04c9c {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1014.827378] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e46c16e9-0970-4e2d-bd13-ddbd7fd76b6a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.833400] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294177, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.838428] env[63021]: DEBUG oslo_vmware.api [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1014.838428] env[63021]: value = "task-1294178" [ 1014.838428] env[63021]: _type = "Task" [ 1014.838428] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.845931] env[63021]: DEBUG oslo_vmware.api [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294178, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.150035] env[63021]: DEBUG oslo_vmware.api [None req-8645ea1e-dd6a-47bb-be9d-3069b9df0403 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294172, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.233850] env[63021]: DEBUG nova.compute.manager [req-6c820ca1-4f6c-427f-84fc-7b437a625b0a req-184ceb9f-b512-41bd-b142-05c214979303 service nova] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Received event network-vif-plugged-1bb789b9-f77e-4e8c-a00b-c513af834d07 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1015.234127] env[63021]: DEBUG oslo_concurrency.lockutils [req-6c820ca1-4f6c-427f-84fc-7b437a625b0a req-184ceb9f-b512-41bd-b142-05c214979303 service nova] Acquiring lock "8d9387b8-6a84-4ada-8ac7-4ffd922b675a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.234312] env[63021]: DEBUG oslo_concurrency.lockutils [req-6c820ca1-4f6c-427f-84fc-7b437a625b0a req-184ceb9f-b512-41bd-b142-05c214979303 service nova] Lock "8d9387b8-6a84-4ada-8ac7-4ffd922b675a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.234479] env[63021]: DEBUG oslo_concurrency.lockutils [req-6c820ca1-4f6c-427f-84fc-7b437a625b0a req-184ceb9f-b512-41bd-b142-05c214979303 service nova] Lock "8d9387b8-6a84-4ada-8ac7-4ffd922b675a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.234683] env[63021]: DEBUG nova.compute.manager [req-6c820ca1-4f6c-427f-84fc-7b437a625b0a req-184ceb9f-b512-41bd-b142-05c214979303 service nova] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] No waiting events found dispatching network-vif-plugged-1bb789b9-f77e-4e8c-a00b-c513af834d07 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1015.234893] env[63021]: WARNING nova.compute.manager [req-6c820ca1-4f6c-427f-84fc-7b437a625b0a req-184ceb9f-b512-41bd-b142-05c214979303 service nova] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Received unexpected event network-vif-plugged-1bb789b9-f77e-4e8c-a00b-c513af834d07 for instance with vm_state building and task_state spawning. [ 1015.268401] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f0ca5997-0f12-4ea9-be55-4f216727adf5 tempest-AttachVolumeShelveTestJSON-999364130 tempest-AttachVolumeShelveTestJSON-999364130-project-member] Lock "45c9e2be-eb19-4b83-b280-c9eeaddfccba" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.532s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.293020] env[63021]: INFO nova.compute.manager [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Took 16.30 seconds to build instance. [ 1015.330774] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294177, 'name': CreateVM_Task, 'duration_secs': 0.368021} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.330954] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1015.331633] env[63021]: DEBUG oslo_concurrency.lockutils [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.331802] env[63021]: DEBUG oslo_concurrency.lockutils [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.332204] env[63021]: DEBUG oslo_concurrency.lockutils [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1015.332553] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0ef6d4d-3f68-4fe4-91b8-e784e02a9921 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.337037] env[63021]: DEBUG oslo_vmware.api [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1015.337037] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5257aaaf-10a9-208c-b971-5769180eb343" [ 1015.337037] env[63021]: _type = "Task" [ 1015.337037] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.348432] env[63021]: DEBUG oslo_vmware.api [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5257aaaf-10a9-208c-b971-5769180eb343, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.351913] env[63021]: DEBUG oslo_vmware.api [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294178, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.216449} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.352158] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1015.352339] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Deleted contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1015.352514] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1015.352703] env[63021]: INFO nova.compute.manager [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1015.352959] env[63021]: DEBUG oslo.service.loopingcall [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1015.353165] env[63021]: DEBUG nova.compute.manager [-] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1015.353447] env[63021]: DEBUG nova.network.neutron [-] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1015.469957] env[63021]: DEBUG nova.network.neutron [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Successfully updated port: 1bb789b9-f77e-4e8c-a00b-c513af834d07 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1015.652115] env[63021]: DEBUG oslo_vmware.api [None req-8645ea1e-dd6a-47bb-be9d-3069b9df0403 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294172, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.735341] env[63021]: DEBUG nova.compute.manager [req-59bf12f5-a82c-4026-9e73-c56f01e69392 req-1dcaf3ea-7498-4c2f-b51b-402e65304780 service nova] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Received event network-vif-deleted-2df835dc-fa26-45ab-9d30-58f85c9e3331 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1015.735341] env[63021]: INFO nova.compute.manager [req-59bf12f5-a82c-4026-9e73-c56f01e69392 req-1dcaf3ea-7498-4c2f-b51b-402e65304780 service nova] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Neutron deleted interface 2df835dc-fa26-45ab-9d30-58f85c9e3331; detaching it from the instance and deleting it from the info cache [ 1015.735341] env[63021]: DEBUG nova.network.neutron [req-59bf12f5-a82c-4026-9e73-c56f01e69392 req-1dcaf3ea-7498-4c2f-b51b-402e65304780 service nova] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.795697] env[63021]: DEBUG oslo_concurrency.lockutils [None req-31c440c4-c62e-4963-9a70-21e7df2416ec tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "07510760-13ff-4797-86ca-86b384aacdfa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.808s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.848405] env[63021]: DEBUG oslo_vmware.api [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5257aaaf-10a9-208c-b971-5769180eb343, 'name': SearchDatastore_Task, 'duration_secs': 0.040655} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.848759] env[63021]: DEBUG oslo_concurrency.lockutils [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.849040] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1015.850407] env[63021]: DEBUG oslo_concurrency.lockutils [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.850407] env[63021]: DEBUG oslo_concurrency.lockutils [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.850407] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1015.850738] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-38c3a2ff-f2e7-4ef8-b047-4a4b34ade5ff {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.862102] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1015.862360] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1015.863714] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-900598fa-c0fd-408f-a0ea-24de7704791e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.871489] env[63021]: DEBUG oslo_vmware.api [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1015.871489] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52149c5f-57bd-c38e-d3d6-e3a32d996996" [ 1015.871489] env[63021]: _type = "Task" [ 1015.871489] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.882441] env[63021]: DEBUG oslo_vmware.api [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52149c5f-57bd-c38e-d3d6-e3a32d996996, 'name': SearchDatastore_Task, 'duration_secs': 0.009031} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.883258] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9fb4c52d-ba1c-4eb3-8ee9-c2b21f7a346d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.888948] env[63021]: DEBUG oslo_vmware.api [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1015.888948] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5219b8f8-a787-1d50-af44-facfef9e6bba" [ 1015.888948] env[63021]: _type = "Task" [ 1015.888948] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.897152] env[63021]: DEBUG oslo_vmware.api [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5219b8f8-a787-1d50-af44-facfef9e6bba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.975255] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquiring lock "refresh_cache-8d9387b8-6a84-4ada-8ac7-4ffd922b675a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.975255] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquired lock "refresh_cache-8d9387b8-6a84-4ada-8ac7-4ffd922b675a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.975255] env[63021]: DEBUG nova.network.neutron [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1016.151021] env[63021]: DEBUG oslo_vmware.api [None req-8645ea1e-dd6a-47bb-be9d-3069b9df0403 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294172, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.206583] env[63021]: DEBUG nova.network.neutron [-] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.237682] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4661d14c-f5f3-493c-9830-95b07b7a12e4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.249892] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2a95261-6d21-4f98-879e-2fb261dacbaa {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.275113] env[63021]: DEBUG nova.compute.manager [req-59bf12f5-a82c-4026-9e73-c56f01e69392 req-1dcaf3ea-7498-4c2f-b51b-402e65304780 service nova] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Detach interface failed, port_id=2df835dc-fa26-45ab-9d30-58f85c9e3331, reason: Instance 023c0891-2b08-48bb-9754-408ce7d04c9c could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1016.279381] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "d03c47da-7e73-4474-ad9b-07d27629dcae" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.279542] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "d03c47da-7e73-4474-ad9b-07d27629dcae" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.358777] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d5b4bdb9-b11a-401c-a6e3-c88143b9f441 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "07510760-13ff-4797-86ca-86b384aacdfa" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.359097] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d5b4bdb9-b11a-401c-a6e3-c88143b9f441 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "07510760-13ff-4797-86ca-86b384aacdfa" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.359288] env[63021]: DEBUG nova.compute.manager [None req-d5b4bdb9-b11a-401c-a6e3-c88143b9f441 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1016.360474] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b70a66ac-dd96-49f5-9416-007d37e14f85 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.367188] env[63021]: DEBUG nova.compute.manager [None req-d5b4bdb9-b11a-401c-a6e3-c88143b9f441 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63021) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1016.367750] env[63021]: DEBUG nova.objects.instance [None req-d5b4bdb9-b11a-401c-a6e3-c88143b9f441 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lazy-loading 'flavor' on Instance uuid 07510760-13ff-4797-86ca-86b384aacdfa {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1016.399220] env[63021]: DEBUG oslo_vmware.api [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5219b8f8-a787-1d50-af44-facfef9e6bba, 'name': SearchDatastore_Task, 'duration_secs': 0.009438} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.399501] env[63021]: DEBUG oslo_concurrency.lockutils [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.399761] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 67aed8eb-d08c-47cc-8558-5c5a0f058d6e/67aed8eb-d08c-47cc-8558-5c5a0f058d6e.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1016.400043] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-46b190b1-0ebe-4629-b59e-b9a6cda514e7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.406244] env[63021]: DEBUG oslo_vmware.api [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1016.406244] env[63021]: value = "task-1294179" [ 1016.406244] env[63021]: _type = "Task" [ 1016.406244] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.414321] env[63021]: DEBUG oslo_vmware.api [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294179, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.507030] env[63021]: DEBUG nova.network.neutron [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1016.650256] env[63021]: DEBUG nova.network.neutron [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Updating instance_info_cache with network_info: [{"id": "1bb789b9-f77e-4e8c-a00b-c513af834d07", "address": "fa:16:3e:41:22:b1", "network": {"id": "7258df3e-b1eb-4da6-ab52-5dadd2f87b51", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1084458843-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5661c44b534b4a7e9452e914fc80d66e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4f04accd-4a9f-41a1-b327-c96095e1ade4", "external-id": "nsx-vlan-transportzone-947", "segmentation_id": 947, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1bb789b9-f7", "ovs_interfaceid": "1bb789b9-f77e-4e8c-a00b-c513af834d07", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.654604] env[63021]: DEBUG oslo_vmware.api [None req-8645ea1e-dd6a-47bb-be9d-3069b9df0403 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294172, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.709836] env[63021]: INFO nova.compute.manager [-] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Took 1.36 seconds to deallocate network for instance. [ 1016.781922] env[63021]: DEBUG nova.compute.manager [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1016.872957] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5b4bdb9-b11a-401c-a6e3-c88143b9f441 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1016.873267] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6b369d21-0adb-4bf6-91f9-7c3637305130 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.879378] env[63021]: DEBUG oslo_vmware.api [None req-d5b4bdb9-b11a-401c-a6e3-c88143b9f441 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1016.879378] env[63021]: value = "task-1294180" [ 1016.879378] env[63021]: _type = "Task" [ 1016.879378] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.887536] env[63021]: DEBUG oslo_vmware.api [None req-d5b4bdb9-b11a-401c-a6e3-c88143b9f441 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294180, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.915273] env[63021]: DEBUG oslo_vmware.api [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294179, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.465287} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.915530] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 67aed8eb-d08c-47cc-8558-5c5a0f058d6e/67aed8eb-d08c-47cc-8558-5c5a0f058d6e.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1016.915745] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1016.915994] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9e91f8dc-b959-46be-9ede-6fdc93fec375 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.921736] env[63021]: DEBUG oslo_vmware.api [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1016.921736] env[63021]: value = "task-1294181" [ 1016.921736] env[63021]: _type = "Task" [ 1016.921736] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.928534] env[63021]: DEBUG oslo_vmware.api [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294181, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.153088] env[63021]: DEBUG oslo_vmware.api [None req-8645ea1e-dd6a-47bb-be9d-3069b9df0403 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294172, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.155681] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Releasing lock "refresh_cache-8d9387b8-6a84-4ada-8ac7-4ffd922b675a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.155982] env[63021]: DEBUG nova.compute.manager [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Instance network_info: |[{"id": "1bb789b9-f77e-4e8c-a00b-c513af834d07", "address": "fa:16:3e:41:22:b1", "network": {"id": "7258df3e-b1eb-4da6-ab52-5dadd2f87b51", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1084458843-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5661c44b534b4a7e9452e914fc80d66e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4f04accd-4a9f-41a1-b327-c96095e1ade4", "external-id": "nsx-vlan-transportzone-947", "segmentation_id": 947, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1bb789b9-f7", "ovs_interfaceid": "1bb789b9-f77e-4e8c-a00b-c513af834d07", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1017.156384] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:41:22:b1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4f04accd-4a9f-41a1-b327-c96095e1ade4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1bb789b9-f77e-4e8c-a00b-c513af834d07', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1017.164474] env[63021]: DEBUG oslo.service.loopingcall [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1017.164724] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1017.165065] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dddfe94c-a44a-4776-ab2d-798f1d2117ec {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.184024] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1017.184024] env[63021]: value = "task-1294182" [ 1017.184024] env[63021]: _type = "Task" [ 1017.184024] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.194135] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294182, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.216995] env[63021]: DEBUG oslo_concurrency.lockutils [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.217284] env[63021]: DEBUG oslo_concurrency.lockutils [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.217551] env[63021]: DEBUG nova.objects.instance [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lazy-loading 'resources' on Instance uuid 023c0891-2b08-48bb-9754-408ce7d04c9c {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1017.272113] env[63021]: DEBUG nova.compute.manager [req-45fb5a67-6c5c-4314-b917-93c7cade1dcd req-635177d7-de4e-4a68-ae92-758f12ec5d34 service nova] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Received event network-changed-1bb789b9-f77e-4e8c-a00b-c513af834d07 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1017.272113] env[63021]: DEBUG nova.compute.manager [req-45fb5a67-6c5c-4314-b917-93c7cade1dcd req-635177d7-de4e-4a68-ae92-758f12ec5d34 service nova] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Refreshing instance network info cache due to event network-changed-1bb789b9-f77e-4e8c-a00b-c513af834d07. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1017.272113] env[63021]: DEBUG oslo_concurrency.lockutils [req-45fb5a67-6c5c-4314-b917-93c7cade1dcd req-635177d7-de4e-4a68-ae92-758f12ec5d34 service nova] Acquiring lock "refresh_cache-8d9387b8-6a84-4ada-8ac7-4ffd922b675a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.272259] env[63021]: DEBUG oslo_concurrency.lockutils [req-45fb5a67-6c5c-4314-b917-93c7cade1dcd req-635177d7-de4e-4a68-ae92-758f12ec5d34 service nova] Acquired lock "refresh_cache-8d9387b8-6a84-4ada-8ac7-4ffd922b675a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.272411] env[63021]: DEBUG nova.network.neutron [req-45fb5a67-6c5c-4314-b917-93c7cade1dcd req-635177d7-de4e-4a68-ae92-758f12ec5d34 service nova] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Refreshing network info cache for port 1bb789b9-f77e-4e8c-a00b-c513af834d07 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1017.306209] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.389975] env[63021]: DEBUG oslo_vmware.api [None req-d5b4bdb9-b11a-401c-a6e3-c88143b9f441 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294180, 'name': PowerOffVM_Task, 'duration_secs': 0.177609} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.390276] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5b4bdb9-b11a-401c-a6e3-c88143b9f441 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1017.390455] env[63021]: DEBUG nova.compute.manager [None req-d5b4bdb9-b11a-401c-a6e3-c88143b9f441 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1017.391335] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb371ad9-3452-4803-8b96-59a070b8d188 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.430995] env[63021]: DEBUG oslo_vmware.api [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294181, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067668} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.431343] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1017.432123] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf14d451-f209-4609-a747-52abba733484 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.456509] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] 67aed8eb-d08c-47cc-8558-5c5a0f058d6e/67aed8eb-d08c-47cc-8558-5c5a0f058d6e.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1017.457093] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b663139a-67e8-4e96-bbd9-d0ad6caeeefe {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.476317] env[63021]: DEBUG oslo_vmware.api [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1017.476317] env[63021]: value = "task-1294183" [ 1017.476317] env[63021]: _type = "Task" [ 1017.476317] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.485343] env[63021]: DEBUG oslo_vmware.api [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294183, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.655797] env[63021]: DEBUG oslo_vmware.api [None req-8645ea1e-dd6a-47bb-be9d-3069b9df0403 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294172, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.694746] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294182, 'name': CreateVM_Task, 'duration_secs': 0.333653} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.694746] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1017.696267] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.696267] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.696267] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1017.696267] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-278e9107-f167-4478-a53d-d1c2adad5f9e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.701251] env[63021]: DEBUG oslo_vmware.api [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 1017.701251] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5265a7e2-608e-7ef0-51f1-5ac559879fd4" [ 1017.701251] env[63021]: _type = "Task" [ 1017.701251] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.708998] env[63021]: DEBUG oslo_vmware.api [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5265a7e2-608e-7ef0-51f1-5ac559879fd4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.903643] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3a7e947-0fd0-4cf7-a222-630ac7677d21 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.906670] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d5b4bdb9-b11a-401c-a6e3-c88143b9f441 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "07510760-13ff-4797-86ca-86b384aacdfa" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.548s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.912184] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c7fffce-07cf-498a-80a3-3a2632369a45 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.945176] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb32cad3-f211-4e82-8129-034ee655e52a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.953048] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-400d3ee7-b5e6-4c51-9179-05cd41e70dac {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.968178] env[63021]: DEBUG nova.compute.provider_tree [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1017.985372] env[63021]: DEBUG oslo_vmware.api [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294183, 'name': ReconfigVM_Task, 'duration_secs': 0.310961} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.985710] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Reconfigured VM instance instance-0000005e to attach disk [datastore2] 67aed8eb-d08c-47cc-8558-5c5a0f058d6e/67aed8eb-d08c-47cc-8558-5c5a0f058d6e.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1017.986340] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-583baa6d-ecf9-4535-b332-458f8ad85fa3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.994087] env[63021]: DEBUG oslo_vmware.api [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1017.994087] env[63021]: value = "task-1294184" [ 1017.994087] env[63021]: _type = "Task" [ 1017.994087] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.003864] env[63021]: DEBUG oslo_vmware.api [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294184, 'name': Rename_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.155215] env[63021]: DEBUG oslo_vmware.api [None req-8645ea1e-dd6a-47bb-be9d-3069b9df0403 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294172, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.211017] env[63021]: DEBUG oslo_vmware.api [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5265a7e2-608e-7ef0-51f1-5ac559879fd4, 'name': SearchDatastore_Task, 'duration_secs': 0.009602} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.211368] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.211605] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1018.211869] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.212035] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.212221] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1018.212495] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-441f2790-c858-4cf2-9dd3-e7e4c6b66bef {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.220911] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1018.221134] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1018.221795] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da32280b-3e3f-4aad-b585-5296589bf1ae {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.227240] env[63021]: DEBUG oslo_vmware.api [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 1018.227240] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52bdbde9-ee68-c85e-48e4-4b67ea7ff191" [ 1018.227240] env[63021]: _type = "Task" [ 1018.227240] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.234959] env[63021]: DEBUG oslo_vmware.api [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52bdbde9-ee68-c85e-48e4-4b67ea7ff191, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.246957] env[63021]: DEBUG nova.network.neutron [req-45fb5a67-6c5c-4314-b917-93c7cade1dcd req-635177d7-de4e-4a68-ae92-758f12ec5d34 service nova] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Updated VIF entry in instance network info cache for port 1bb789b9-f77e-4e8c-a00b-c513af834d07. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1018.247357] env[63021]: DEBUG nova.network.neutron [req-45fb5a67-6c5c-4314-b917-93c7cade1dcd req-635177d7-de4e-4a68-ae92-758f12ec5d34 service nova] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Updating instance_info_cache with network_info: [{"id": "1bb789b9-f77e-4e8c-a00b-c513af834d07", "address": "fa:16:3e:41:22:b1", "network": {"id": "7258df3e-b1eb-4da6-ab52-5dadd2f87b51", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1084458843-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5661c44b534b4a7e9452e914fc80d66e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4f04accd-4a9f-41a1-b327-c96095e1ade4", "external-id": "nsx-vlan-transportzone-947", "segmentation_id": 947, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1bb789b9-f7", "ovs_interfaceid": "1bb789b9-f77e-4e8c-a00b-c513af834d07", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.470959] env[63021]: DEBUG nova.scheduler.client.report [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1018.505529] env[63021]: DEBUG oslo_vmware.api [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294184, 'name': Rename_Task, 'duration_secs': 0.156523} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.505814] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1018.506079] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-66fbe704-cf99-424e-bd8f-3133a2cc582c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.512226] env[63021]: DEBUG oslo_vmware.api [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1018.512226] env[63021]: value = "task-1294185" [ 1018.512226] env[63021]: _type = "Task" [ 1018.512226] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.519793] env[63021]: DEBUG oslo_vmware.api [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294185, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.657580] env[63021]: DEBUG oslo_vmware.api [None req-8645ea1e-dd6a-47bb-be9d-3069b9df0403 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294172, 'name': ReconfigVM_Task, 'duration_secs': 5.776336} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.657580] env[63021]: DEBUG oslo_concurrency.lockutils [None req-8645ea1e-dd6a-47bb-be9d-3069b9df0403 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Releasing lock "6155c098-e4a6-47e6-b343-4a77ca90eb2e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.657925] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-8645ea1e-dd6a-47bb-be9d-3069b9df0403 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Reconfigured VM to detach interface {{(pid=63021) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1018.669928] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b3b3cff2-7134-42ec-89e0-a3a1192740e5 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "67f736d9-82ec-448e-8b72-84fdd704aa8d" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.670196] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b3b3cff2-7134-42ec-89e0-a3a1192740e5 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "67f736d9-82ec-448e-8b72-84fdd704aa8d" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.739438] env[63021]: DEBUG oslo_vmware.api [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52bdbde9-ee68-c85e-48e4-4b67ea7ff191, 'name': SearchDatastore_Task, 'duration_secs': 0.016334} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.740615] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c94e537b-8c98-470a-bfd8-de4d27078037 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.745380] env[63021]: DEBUG oslo_vmware.api [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 1018.745380] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5295aa71-9556-7f3b-2344-579e8f963214" [ 1018.745380] env[63021]: _type = "Task" [ 1018.745380] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.750838] env[63021]: DEBUG oslo_concurrency.lockutils [req-45fb5a67-6c5c-4314-b917-93c7cade1dcd req-635177d7-de4e-4a68-ae92-758f12ec5d34 service nova] Releasing lock "refresh_cache-8d9387b8-6a84-4ada-8ac7-4ffd922b675a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.755564] env[63021]: DEBUG oslo_vmware.api [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5295aa71-9556-7f3b-2344-579e8f963214, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.892037] env[63021]: DEBUG oslo_concurrency.lockutils [None req-606c6549-987d-4df9-8e61-c66d7a750924 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "07510760-13ff-4797-86ca-86b384aacdfa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.892424] env[63021]: DEBUG oslo_concurrency.lockutils [None req-606c6549-987d-4df9-8e61-c66d7a750924 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "07510760-13ff-4797-86ca-86b384aacdfa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.892644] env[63021]: DEBUG oslo_concurrency.lockutils [None req-606c6549-987d-4df9-8e61-c66d7a750924 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "07510760-13ff-4797-86ca-86b384aacdfa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.892971] env[63021]: DEBUG oslo_concurrency.lockutils [None req-606c6549-987d-4df9-8e61-c66d7a750924 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "07510760-13ff-4797-86ca-86b384aacdfa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.893081] env[63021]: DEBUG oslo_concurrency.lockutils [None req-606c6549-987d-4df9-8e61-c66d7a750924 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "07510760-13ff-4797-86ca-86b384aacdfa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.896875] env[63021]: INFO nova.compute.manager [None req-606c6549-987d-4df9-8e61-c66d7a750924 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Terminating instance [ 1018.898797] env[63021]: DEBUG nova.compute.manager [None req-606c6549-987d-4df9-8e61-c66d7a750924 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1018.898998] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-606c6549-987d-4df9-8e61-c66d7a750924 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1018.899877] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24ecdb5b-be4a-47cc-a456-7deb80133bc7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.907765] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-606c6549-987d-4df9-8e61-c66d7a750924 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1018.908079] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9dbdd38c-4d36-4bbc-bf15-18ea09c2bfbf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.967895] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-606c6549-987d-4df9-8e61-c66d7a750924 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1018.968201] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-606c6549-987d-4df9-8e61-c66d7a750924 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Deleting contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1018.968391] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-606c6549-987d-4df9-8e61-c66d7a750924 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Deleting the datastore file [datastore2] 07510760-13ff-4797-86ca-86b384aacdfa {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1018.968684] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8433d5e2-c128-4b0f-9194-277c0b91fbea {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.976111] env[63021]: DEBUG oslo_concurrency.lockutils [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.759s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.978452] env[63021]: DEBUG oslo_vmware.api [None req-606c6549-987d-4df9-8e61-c66d7a750924 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1018.978452] env[63021]: value = "task-1294187" [ 1018.978452] env[63021]: _type = "Task" [ 1018.978452] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.978995] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.673s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.980790] env[63021]: INFO nova.compute.claims [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1018.992198] env[63021]: DEBUG oslo_vmware.api [None req-606c6549-987d-4df9-8e61-c66d7a750924 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294187, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.003812] env[63021]: INFO nova.scheduler.client.report [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Deleted allocations for instance 023c0891-2b08-48bb-9754-408ce7d04c9c [ 1019.021879] env[63021]: DEBUG oslo_vmware.api [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294185, 'name': PowerOnVM_Task, 'duration_secs': 0.448026} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.022176] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1019.022390] env[63021]: DEBUG nova.compute.manager [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1019.023519] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9817dfce-b69b-4764-8175-25b2af1a1f7e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.173774] env[63021]: DEBUG nova.compute.utils [None req-b3b3cff2-7134-42ec-89e0-a3a1192740e5 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1019.257110] env[63021]: DEBUG oslo_vmware.api [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5295aa71-9556-7f3b-2344-579e8f963214, 'name': SearchDatastore_Task, 'duration_secs': 0.015007} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.257358] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.257615] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 8d9387b8-6a84-4ada-8ac7-4ffd922b675a/8d9387b8-6a84-4ada-8ac7-4ffd922b675a.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1019.257926] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-eae0055b-aa30-4da5-8210-f2812e495fb3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.264363] env[63021]: DEBUG oslo_vmware.api [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 1019.264363] env[63021]: value = "task-1294188" [ 1019.264363] env[63021]: _type = "Task" [ 1019.264363] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.272411] env[63021]: DEBUG oslo_vmware.api [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294188, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.493411] env[63021]: DEBUG oslo_vmware.api [None req-606c6549-987d-4df9-8e61-c66d7a750924 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294187, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.256625} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.494217] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-606c6549-987d-4df9-8e61-c66d7a750924 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1019.494434] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-606c6549-987d-4df9-8e61-c66d7a750924 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Deleted contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1019.494634] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-606c6549-987d-4df9-8e61-c66d7a750924 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1019.494767] env[63021]: INFO nova.compute.manager [None req-606c6549-987d-4df9-8e61-c66d7a750924 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Took 0.60 seconds to destroy the instance on the hypervisor. [ 1019.495068] env[63021]: DEBUG oslo.service.loopingcall [None req-606c6549-987d-4df9-8e61-c66d7a750924 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1019.495279] env[63021]: DEBUG nova.compute.manager [-] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1019.495374] env[63021]: DEBUG nova.network.neutron [-] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1019.513656] env[63021]: DEBUG oslo_concurrency.lockutils [None req-80a29b86-d4d3-451c-b346-c9bafc8f9528 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "023c0891-2b08-48bb-9754-408ce7d04c9c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.353s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.540646] env[63021]: DEBUG oslo_concurrency.lockutils [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.677153] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b3b3cff2-7134-42ec-89e0-a3a1192740e5 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "67f736d9-82ec-448e-8b72-84fdd704aa8d" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.778100] env[63021]: DEBUG oslo_vmware.api [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294188, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.850718] env[63021]: DEBUG nova.compute.manager [req-08e93dae-9af2-4bee-971a-9056d23c4511 req-88aa8722-9090-4864-b1d8-849c89419457 service nova] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Received event network-vif-deleted-4a29562f-684a-4738-98b0-f1281b0bc4c3 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1019.850988] env[63021]: INFO nova.compute.manager [req-08e93dae-9af2-4bee-971a-9056d23c4511 req-88aa8722-9090-4864-b1d8-849c89419457 service nova] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Neutron deleted interface 4a29562f-684a-4738-98b0-f1281b0bc4c3; detaching it from the instance and deleting it from the info cache [ 1019.851167] env[63021]: DEBUG nova.network.neutron [req-08e93dae-9af2-4bee-971a-9056d23c4511 req-88aa8722-9090-4864-b1d8-849c89419457 service nova] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.077369] env[63021]: DEBUG oslo_concurrency.lockutils [None req-8645ea1e-dd6a-47bb-be9d-3069b9df0403 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "refresh_cache-6155c098-e4a6-47e6-b343-4a77ca90eb2e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.077739] env[63021]: DEBUG oslo_concurrency.lockutils [None req-8645ea1e-dd6a-47bb-be9d-3069b9df0403 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquired lock "refresh_cache-6155c098-e4a6-47e6-b343-4a77ca90eb2e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.077739] env[63021]: DEBUG nova.network.neutron [None req-8645ea1e-dd6a-47bb-be9d-3069b9df0403 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1020.143183] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fccfec3-55ef-4cb0-a884-316a10437928 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.150681] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6906314-724c-43b8-842d-ed662bf0cdf6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.192804] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ed382e7-9b67-4918-ab95-8a790d37cd30 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.200404] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92073744-dd23-4ab8-add0-78d942a62b7a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.214746] env[63021]: DEBUG nova.compute.provider_tree [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1020.271721] env[63021]: DEBUG nova.network.neutron [-] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.279231] env[63021]: DEBUG oslo_vmware.api [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294188, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.690203} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.279231] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 8d9387b8-6a84-4ada-8ac7-4ffd922b675a/8d9387b8-6a84-4ada-8ac7-4ffd922b675a.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1020.279443] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1020.279590] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-114ed2c7-d9c1-4035-a493-0c8f22cd266b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.287089] env[63021]: DEBUG oslo_vmware.api [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 1020.287089] env[63021]: value = "task-1294189" [ 1020.287089] env[63021]: _type = "Task" [ 1020.287089] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.294972] env[63021]: DEBUG oslo_vmware.api [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294189, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.354034] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-523baf00-568c-4438-9710-6abf445bd33d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.363546] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3ad32b3-9207-42d4-bc1c-a77f6b0cae60 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.392509] env[63021]: DEBUG nova.compute.manager [req-08e93dae-9af2-4bee-971a-9056d23c4511 req-88aa8722-9090-4864-b1d8-849c89419457 service nova] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Detach interface failed, port_id=4a29562f-684a-4738-98b0-f1281b0bc4c3, reason: Instance 07510760-13ff-4797-86ca-86b384aacdfa could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1020.422599] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "6155c098-e4a6-47e6-b343-4a77ca90eb2e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.423402] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "6155c098-e4a6-47e6-b343-4a77ca90eb2e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.423402] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "6155c098-e4a6-47e6-b343-4a77ca90eb2e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.423402] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "6155c098-e4a6-47e6-b343-4a77ca90eb2e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.423402] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "6155c098-e4a6-47e6-b343-4a77ca90eb2e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.425709] env[63021]: INFO nova.compute.manager [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Terminating instance [ 1020.428010] env[63021]: DEBUG nova.compute.manager [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1020.428217] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1020.429028] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6847b6a7-a4de-4a34-83cf-77f3008aeb83 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.435781] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1020.436013] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ca531743-589c-4b6f-b418-992be993af08 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.441999] env[63021]: DEBUG oslo_vmware.api [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 1020.441999] env[63021]: value = "task-1294190" [ 1020.441999] env[63021]: _type = "Task" [ 1020.441999] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.449536] env[63021]: DEBUG oslo_vmware.api [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294190, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.717402] env[63021]: DEBUG nova.scheduler.client.report [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1020.739587] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b3b3cff2-7134-42ec-89e0-a3a1192740e5 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "67f736d9-82ec-448e-8b72-84fdd704aa8d" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.740051] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b3b3cff2-7134-42ec-89e0-a3a1192740e5 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "67f736d9-82ec-448e-8b72-84fdd704aa8d" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.740115] env[63021]: INFO nova.compute.manager [None req-b3b3cff2-7134-42ec-89e0-a3a1192740e5 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Attaching volume 236a8786-fa3f-411d-a486-ddd5f2950147 to /dev/sdb [ 1020.771526] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2311b4bc-b020-4f47-8ea1-28e2002e3c5e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.776553] env[63021]: INFO nova.compute.manager [-] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Took 1.28 seconds to deallocate network for instance. [ 1020.783159] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0514c134-f81e-40c5-aee9-09a65ccbf79c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.796834] env[63021]: DEBUG nova.virt.block_device [None req-b3b3cff2-7134-42ec-89e0-a3a1192740e5 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Updating existing volume attachment record: adf1dbf4-635d-48ca-8a7b-e120b022933f {{(pid=63021) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1020.801652] env[63021]: DEBUG oslo_vmware.api [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294189, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.108688} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.801910] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1020.803263] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8c9d7c2-595f-41ec-9e95-2b54a6c6c688 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.824270] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] 8d9387b8-6a84-4ada-8ac7-4ffd922b675a/8d9387b8-6a84-4ada-8ac7-4ffd922b675a.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1020.825252] env[63021]: INFO nova.network.neutron [None req-8645ea1e-dd6a-47bb-be9d-3069b9df0403 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Port 01e26b49-58a2-411d-b025-b1ca941026c4 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1020.825596] env[63021]: DEBUG nova.network.neutron [None req-8645ea1e-dd6a-47bb-be9d-3069b9df0403 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Updating instance_info_cache with network_info: [{"id": "3752044a-9145-469d-8edb-6255d7d0f928", "address": "fa:16:3e:fe:57:1f", "network": {"id": "2c4f9111-2f39-4bd0-95bc-09b96d535ea3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-402684019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58693d1e991249f28914a39761cc2627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3752044a-91", "ovs_interfaceid": "3752044a-9145-469d-8edb-6255d7d0f928", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.827554] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5bf3a34f-3c01-4cbd-a5c3-08bd458b23fb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.846243] env[63021]: DEBUG oslo_vmware.api [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 1020.846243] env[63021]: value = "task-1294191" [ 1020.846243] env[63021]: _type = "Task" [ 1020.846243] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.854279] env[63021]: DEBUG oslo_vmware.api [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294191, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.881672] env[63021]: DEBUG oslo_concurrency.lockutils [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "67aed8eb-d08c-47cc-8558-5c5a0f058d6e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.881948] env[63021]: DEBUG oslo_concurrency.lockutils [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "67aed8eb-d08c-47cc-8558-5c5a0f058d6e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.882183] env[63021]: DEBUG oslo_concurrency.lockutils [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "67aed8eb-d08c-47cc-8558-5c5a0f058d6e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.882373] env[63021]: DEBUG oslo_concurrency.lockutils [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "67aed8eb-d08c-47cc-8558-5c5a0f058d6e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.882555] env[63021]: DEBUG oslo_concurrency.lockutils [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "67aed8eb-d08c-47cc-8558-5c5a0f058d6e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.884804] env[63021]: INFO nova.compute.manager [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Terminating instance [ 1020.886872] env[63021]: DEBUG nova.compute.manager [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1020.887127] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1020.888050] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-204f3c06-d7ef-42ca-9220-04b9cc6bc84d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.895141] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1020.895342] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cb556ef8-996a-4c5f-9b28-942ab3c85238 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.900429] env[63021]: DEBUG oslo_vmware.api [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1020.900429] env[63021]: value = "task-1294192" [ 1020.900429] env[63021]: _type = "Task" [ 1020.900429] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.908104] env[63021]: DEBUG oslo_vmware.api [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294192, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.951646] env[63021]: DEBUG oslo_vmware.api [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294190, 'name': PowerOffVM_Task, 'duration_secs': 0.234915} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.951854] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1020.952109] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1020.952397] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-980274ec-134a-48dc-b09f-00f13482ebdb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.019121] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1021.019444] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Deleting contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1021.019636] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Deleting the datastore file [datastore2] 6155c098-e4a6-47e6-b343-4a77ca90eb2e {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1021.019905] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f08bbacd-5655-4e88-bbf8-f0ce4e5969ac {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.027938] env[63021]: DEBUG oslo_vmware.api [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 1021.027938] env[63021]: value = "task-1294197" [ 1021.027938] env[63021]: _type = "Task" [ 1021.027938] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.035306] env[63021]: DEBUG oslo_vmware.api [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294197, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.223740] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.245s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.224496] env[63021]: DEBUG nova.compute.manager [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1021.228100] env[63021]: DEBUG oslo_concurrency.lockutils [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 1.687s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.228359] env[63021]: DEBUG nova.objects.instance [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63021) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1021.284387] env[63021]: DEBUG oslo_concurrency.lockutils [None req-606c6549-987d-4df9-8e61-c66d7a750924 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.327658] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "63d5b0e4-0d61-4fa7-b193-2eaa171fc461" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.327940] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "63d5b0e4-0d61-4fa7-b193-2eaa171fc461" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.342632] env[63021]: DEBUG oslo_concurrency.lockutils [None req-8645ea1e-dd6a-47bb-be9d-3069b9df0403 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Releasing lock "refresh_cache-6155c098-e4a6-47e6-b343-4a77ca90eb2e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.357309] env[63021]: DEBUG oslo_vmware.api [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294191, 'name': ReconfigVM_Task, 'duration_secs': 0.281378} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.357587] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Reconfigured VM instance instance-00000061 to attach disk [datastore2] 8d9387b8-6a84-4ada-8ac7-4ffd922b675a/8d9387b8-6a84-4ada-8ac7-4ffd922b675a.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1021.358240] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e841ccbb-1fde-45de-9cbe-df97c893a8f6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.364944] env[63021]: DEBUG oslo_vmware.api [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 1021.364944] env[63021]: value = "task-1294198" [ 1021.364944] env[63021]: _type = "Task" [ 1021.364944] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.372923] env[63021]: DEBUG oslo_vmware.api [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294198, 'name': Rename_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.411937] env[63021]: DEBUG oslo_vmware.api [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294192, 'name': PowerOffVM_Task, 'duration_secs': 0.284958} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.412223] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1021.412387] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1021.412649] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-341ec9f4-0be4-4333-92f3-34f4e136d7bc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.480648] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1021.481046] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Deleting contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1021.481296] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Deleting the datastore file [datastore2] 67aed8eb-d08c-47cc-8558-5c5a0f058d6e {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1021.481574] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6ca32b39-0a36-4f05-9b25-01c4a31fa72d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.488268] env[63021]: DEBUG oslo_vmware.api [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1021.488268] env[63021]: value = "task-1294200" [ 1021.488268] env[63021]: _type = "Task" [ 1021.488268] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.496806] env[63021]: DEBUG oslo_vmware.api [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294200, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.538673] env[63021]: DEBUG oslo_vmware.api [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294197, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.170762} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.539015] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1021.539294] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Deleted contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1021.539507] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1021.539684] env[63021]: INFO nova.compute.manager [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1021.539930] env[63021]: DEBUG oslo.service.loopingcall [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1021.540195] env[63021]: DEBUG nova.compute.manager [-] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1021.540332] env[63021]: DEBUG nova.network.neutron [-] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1021.733712] env[63021]: DEBUG nova.compute.utils [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1021.738346] env[63021]: DEBUG nova.compute.manager [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1021.738520] env[63021]: DEBUG nova.network.neutron [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1021.790558] env[63021]: DEBUG nova.policy [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5431727705ff4c5483cd9a5b9cd104f6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '547d9a3099b94ff488ef6929ae591bac', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 1021.830447] env[63021]: DEBUG nova.compute.manager [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1021.846543] env[63021]: DEBUG oslo_concurrency.lockutils [None req-8645ea1e-dd6a-47bb-be9d-3069b9df0403 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "interface-6155c098-e4a6-47e6-b343-4a77ca90eb2e-01e26b49-58a2-411d-b025-b1ca941026c4" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.893s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.876608] env[63021]: DEBUG oslo_vmware.api [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294198, 'name': Rename_Task, 'duration_secs': 0.203835} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.877160] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1021.877246] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b2327fc5-1108-4bf6-bd85-c69c8c1c0f1a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.883110] env[63021]: DEBUG oslo_vmware.api [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 1021.883110] env[63021]: value = "task-1294201" [ 1021.883110] env[63021]: _type = "Task" [ 1021.883110] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.892858] env[63021]: DEBUG oslo_vmware.api [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294201, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.998883] env[63021]: DEBUG oslo_vmware.api [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294200, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.260215} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.999262] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1021.999500] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Deleted contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1021.999730] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1021.999936] env[63021]: INFO nova.compute.manager [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1022.000204] env[63021]: DEBUG oslo.service.loopingcall [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1022.000447] env[63021]: DEBUG nova.compute.manager [-] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1022.000523] env[63021]: DEBUG nova.network.neutron [-] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1022.072373] env[63021]: DEBUG nova.network.neutron [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Successfully created port: 43370cd4-afaf-437d-8079-cf9ad90682aa {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1022.239627] env[63021]: DEBUG nova.compute.manager [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1022.245620] env[63021]: DEBUG oslo_concurrency.lockutils [None req-da4923c4-8f50-49ae-b2c0-9e6a6d0407f6 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.245620] env[63021]: DEBUG oslo_concurrency.lockutils [None req-606c6549-987d-4df9-8e61-c66d7a750924 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.961s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.245620] env[63021]: DEBUG nova.objects.instance [None req-606c6549-987d-4df9-8e61-c66d7a750924 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lazy-loading 'resources' on Instance uuid 07510760-13ff-4797-86ca-86b384aacdfa {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1022.369436] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.396158] env[63021]: DEBUG oslo_vmware.api [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294201, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.416588] env[63021]: DEBUG nova.compute.manager [req-19d37f9f-d532-4f99-87c7-666c9ecb39f3 req-28ca00c3-1ccf-46c0-9a8b-dea44d4d7d18 service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Received event network-vif-deleted-3752044a-9145-469d-8edb-6255d7d0f928 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1022.416931] env[63021]: INFO nova.compute.manager [req-19d37f9f-d532-4f99-87c7-666c9ecb39f3 req-28ca00c3-1ccf-46c0-9a8b-dea44d4d7d18 service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Neutron deleted interface 3752044a-9145-469d-8edb-6255d7d0f928; detaching it from the instance and deleting it from the info cache [ 1022.417194] env[63021]: DEBUG nova.network.neutron [req-19d37f9f-d532-4f99-87c7-666c9ecb39f3 req-28ca00c3-1ccf-46c0-9a8b-dea44d4d7d18 service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.835538] env[63021]: DEBUG nova.network.neutron [-] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.884100] env[63021]: DEBUG nova.network.neutron [-] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.894649] env[63021]: DEBUG oslo_vmware.api [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294201, 'name': PowerOnVM_Task, 'duration_secs': 0.524536} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.895319] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1022.895536] env[63021]: INFO nova.compute.manager [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Took 8.23 seconds to spawn the instance on the hypervisor. [ 1022.895862] env[63021]: DEBUG nova.compute.manager [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1022.896511] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc0175e4-63f8-404b-9329-d2d824e6e632 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.907034] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a30db868-aafb-4421-b5d5-2d7ad32c5b1b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.913875] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f7437ca-bff6-413b-9773-92fab809d6ab {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.945109] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a08c5f2b-bd79-430e-b28a-28e7fb380a09 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.947493] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-415ffb64-39e4-4700-9386-1b527ea150c6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.956521] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e87f3bc8-35c6-4af2-8d70-02688d29b120 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.962952] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a6ac296-7d2a-4835-9d7c-7b7b5fcb1ba1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.981975] env[63021]: DEBUG nova.compute.provider_tree [None req-606c6549-987d-4df9-8e61-c66d7a750924 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1022.991810] env[63021]: DEBUG nova.compute.manager [req-19d37f9f-d532-4f99-87c7-666c9ecb39f3 req-28ca00c3-1ccf-46c0-9a8b-dea44d4d7d18 service nova] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Detach interface failed, port_id=3752044a-9145-469d-8edb-6255d7d0f928, reason: Instance 6155c098-e4a6-47e6-b343-4a77ca90eb2e could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1022.992825] env[63021]: DEBUG nova.scheduler.client.report [None req-606c6549-987d-4df9-8e61-c66d7a750924 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1023.254430] env[63021]: DEBUG nova.compute.manager [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1023.280856] env[63021]: DEBUG nova.virt.hardware [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1023.281125] env[63021]: DEBUG nova.virt.hardware [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1023.281291] env[63021]: DEBUG nova.virt.hardware [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1023.281480] env[63021]: DEBUG nova.virt.hardware [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1023.281619] env[63021]: DEBUG nova.virt.hardware [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1023.281765] env[63021]: DEBUG nova.virt.hardware [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1023.282017] env[63021]: DEBUG nova.virt.hardware [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1023.282195] env[63021]: DEBUG nova.virt.hardware [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1023.282366] env[63021]: DEBUG nova.virt.hardware [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1023.282530] env[63021]: DEBUG nova.virt.hardware [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1023.282708] env[63021]: DEBUG nova.virt.hardware [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1023.283604] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97bf7972-fc39-4d8e-9388-3acd20d95b25 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.291598] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87488d00-cb7f-4fbb-8e2e-fc5f65acecae {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.338271] env[63021]: INFO nova.compute.manager [-] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Took 1.80 seconds to deallocate network for instance. [ 1023.389855] env[63021]: INFO nova.compute.manager [-] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Took 1.39 seconds to deallocate network for instance. [ 1023.416580] env[63021]: INFO nova.compute.manager [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Took 13.64 seconds to build instance. [ 1023.497297] env[63021]: DEBUG oslo_concurrency.lockutils [None req-606c6549-987d-4df9-8e61-c66d7a750924 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.252s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.499977] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.131s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.501810] env[63021]: INFO nova.compute.claims [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1023.521028] env[63021]: INFO nova.scheduler.client.report [None req-606c6549-987d-4df9-8e61-c66d7a750924 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Deleted allocations for instance 07510760-13ff-4797-86ca-86b384aacdfa [ 1023.719719] env[63021]: DEBUG nova.compute.manager [req-5234d866-163d-4a3c-b427-ebd6723f6497 req-fec98480-f1fb-4835-b2c5-0a5a8f7e1f25 service nova] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Received event network-vif-plugged-43370cd4-afaf-437d-8079-cf9ad90682aa {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1023.719719] env[63021]: DEBUG oslo_concurrency.lockutils [req-5234d866-163d-4a3c-b427-ebd6723f6497 req-fec98480-f1fb-4835-b2c5-0a5a8f7e1f25 service nova] Acquiring lock "d03c47da-7e73-4474-ad9b-07d27629dcae-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.719719] env[63021]: DEBUG oslo_concurrency.lockutils [req-5234d866-163d-4a3c-b427-ebd6723f6497 req-fec98480-f1fb-4835-b2c5-0a5a8f7e1f25 service nova] Lock "d03c47da-7e73-4474-ad9b-07d27629dcae-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.719719] env[63021]: DEBUG oslo_concurrency.lockutils [req-5234d866-163d-4a3c-b427-ebd6723f6497 req-fec98480-f1fb-4835-b2c5-0a5a8f7e1f25 service nova] Lock "d03c47da-7e73-4474-ad9b-07d27629dcae-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.719719] env[63021]: DEBUG nova.compute.manager [req-5234d866-163d-4a3c-b427-ebd6723f6497 req-fec98480-f1fb-4835-b2c5-0a5a8f7e1f25 service nova] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] No waiting events found dispatching network-vif-plugged-43370cd4-afaf-437d-8079-cf9ad90682aa {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1023.719719] env[63021]: WARNING nova.compute.manager [req-5234d866-163d-4a3c-b427-ebd6723f6497 req-fec98480-f1fb-4835-b2c5-0a5a8f7e1f25 service nova] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Received unexpected event network-vif-plugged-43370cd4-afaf-437d-8079-cf9ad90682aa for instance with vm_state building and task_state spawning. [ 1023.818760] env[63021]: DEBUG nova.network.neutron [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Successfully updated port: 43370cd4-afaf-437d-8079-cf9ad90682aa {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1023.845105] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.899047] env[63021]: DEBUG oslo_concurrency.lockutils [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.918832] env[63021]: DEBUG oslo_concurrency.lockutils [None req-d1f0b22c-974f-4e88-8950-65615e67e13c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "8d9387b8-6a84-4ada-8ac7-4ffd922b675a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.160s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.029848] env[63021]: DEBUG oslo_concurrency.lockutils [None req-606c6549-987d-4df9-8e61-c66d7a750924 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "07510760-13ff-4797-86ca-86b384aacdfa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.137s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.322748] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "refresh_cache-d03c47da-7e73-4474-ad9b-07d27629dcae" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.322913] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquired lock "refresh_cache-d03c47da-7e73-4474-ad9b-07d27629dcae" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.323081] env[63021]: DEBUG nova.network.neutron [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1024.448735] env[63021]: DEBUG nova.compute.manager [req-d1dc68e0-7773-4ae0-b0a5-dd3faa509ea2 req-8564d181-1b0f-4113-9d2a-fbad7a2504d3 service nova] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Received event network-vif-deleted-371c655a-b09c-48dc-ad0a-da708503e86e {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1024.448735] env[63021]: DEBUG nova.compute.manager [req-d1dc68e0-7773-4ae0-b0a5-dd3faa509ea2 req-8564d181-1b0f-4113-9d2a-fbad7a2504d3 service nova] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Received event network-changed-1bb789b9-f77e-4e8c-a00b-c513af834d07 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1024.449027] env[63021]: DEBUG nova.compute.manager [req-d1dc68e0-7773-4ae0-b0a5-dd3faa509ea2 req-8564d181-1b0f-4113-9d2a-fbad7a2504d3 service nova] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Refreshing instance network info cache due to event network-changed-1bb789b9-f77e-4e8c-a00b-c513af834d07. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1024.449258] env[63021]: DEBUG oslo_concurrency.lockutils [req-d1dc68e0-7773-4ae0-b0a5-dd3faa509ea2 req-8564d181-1b0f-4113-9d2a-fbad7a2504d3 service nova] Acquiring lock "refresh_cache-8d9387b8-6a84-4ada-8ac7-4ffd922b675a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.449405] env[63021]: DEBUG oslo_concurrency.lockutils [req-d1dc68e0-7773-4ae0-b0a5-dd3faa509ea2 req-8564d181-1b0f-4113-9d2a-fbad7a2504d3 service nova] Acquired lock "refresh_cache-8d9387b8-6a84-4ada-8ac7-4ffd922b675a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.449567] env[63021]: DEBUG nova.network.neutron [req-d1dc68e0-7773-4ae0-b0a5-dd3faa509ea2 req-8564d181-1b0f-4113-9d2a-fbad7a2504d3 service nova] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Refreshing network info cache for port 1bb789b9-f77e-4e8c-a00b-c513af834d07 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1024.645203] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a77eeb9-7921-4fe6-97a0-9553671f7986 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.652929] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c5cd7a8-f155-4ef4-9311-daadd38ee658 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.683598] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbf14f45-f909-4e40-ae01-489929e827d7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.691459] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa58ddc1-136d-4fcd-8b08-6b1ce6fc4c9f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.705119] env[63021]: DEBUG nova.compute.provider_tree [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1024.805763] env[63021]: DEBUG oslo_concurrency.lockutils [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "acc38ba0-515d-41d4-81dd-2b4374e033e7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.806145] env[63021]: DEBUG oslo_concurrency.lockutils [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "acc38ba0-515d-41d4-81dd-2b4374e033e7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.864692] env[63021]: DEBUG nova.network.neutron [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1025.022317] env[63021]: DEBUG nova.network.neutron [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Updating instance_info_cache with network_info: [{"id": "43370cd4-afaf-437d-8079-cf9ad90682aa", "address": "fa:16:3e:9b:4c:5d", "network": {"id": "d27bb3cc-fb80-4b03-9288-ed17a0d05140", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-122910456-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "547d9a3099b94ff488ef6929ae591bac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1c797172-a569-458e-aeb0-3f21e589a740", "external-id": "nsx-vlan-transportzone-957", "segmentation_id": 957, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43370cd4-af", "ovs_interfaceid": "43370cd4-afaf-437d-8079-cf9ad90682aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.173642] env[63021]: DEBUG nova.network.neutron [req-d1dc68e0-7773-4ae0-b0a5-dd3faa509ea2 req-8564d181-1b0f-4113-9d2a-fbad7a2504d3 service nova] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Updated VIF entry in instance network info cache for port 1bb789b9-f77e-4e8c-a00b-c513af834d07. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1025.174042] env[63021]: DEBUG nova.network.neutron [req-d1dc68e0-7773-4ae0-b0a5-dd3faa509ea2 req-8564d181-1b0f-4113-9d2a-fbad7a2504d3 service nova] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Updating instance_info_cache with network_info: [{"id": "1bb789b9-f77e-4e8c-a00b-c513af834d07", "address": "fa:16:3e:41:22:b1", "network": {"id": "7258df3e-b1eb-4da6-ab52-5dadd2f87b51", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1084458843-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.226", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5661c44b534b4a7e9452e914fc80d66e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4f04accd-4a9f-41a1-b327-c96095e1ade4", "external-id": "nsx-vlan-transportzone-947", "segmentation_id": 947, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1bb789b9-f7", "ovs_interfaceid": "1bb789b9-f77e-4e8c-a00b-c513af834d07", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.207970] env[63021]: DEBUG nova.scheduler.client.report [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1025.309033] env[63021]: DEBUG nova.compute.manager [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1025.343788] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3b3cff2-7134-42ec-89e0-a3a1192740e5 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Volume attach. Driver type: vmdk {{(pid=63021) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1025.344049] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3b3cff2-7134-42ec-89e0-a3a1192740e5 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277584', 'volume_id': '236a8786-fa3f-411d-a486-ddd5f2950147', 'name': 'volume-236a8786-fa3f-411d-a486-ddd5f2950147', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '67f736d9-82ec-448e-8b72-84fdd704aa8d', 'attached_at': '', 'detached_at': '', 'volume_id': '236a8786-fa3f-411d-a486-ddd5f2950147', 'serial': '236a8786-fa3f-411d-a486-ddd5f2950147'} {{(pid=63021) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1025.344979] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40fa3390-ffc8-42e9-b8fe-a1d9373a5d9f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.362070] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a537b721-ef53-4976-8e48-f798020b04e5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.386488] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3b3cff2-7134-42ec-89e0-a3a1192740e5 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] volume-236a8786-fa3f-411d-a486-ddd5f2950147/volume-236a8786-fa3f-411d-a486-ddd5f2950147.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1025.386744] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd1c4180-d6f8-44a1-aa01-280993729327 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.404016] env[63021]: DEBUG oslo_vmware.api [None req-b3b3cff2-7134-42ec-89e0-a3a1192740e5 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 1025.404016] env[63021]: value = "task-1294203" [ 1025.404016] env[63021]: _type = "Task" [ 1025.404016] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.411625] env[63021]: DEBUG oslo_vmware.api [None req-b3b3cff2-7134-42ec-89e0-a3a1192740e5 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294203, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.527708] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Releasing lock "refresh_cache-d03c47da-7e73-4474-ad9b-07d27629dcae" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.528066] env[63021]: DEBUG nova.compute.manager [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Instance network_info: |[{"id": "43370cd4-afaf-437d-8079-cf9ad90682aa", "address": "fa:16:3e:9b:4c:5d", "network": {"id": "d27bb3cc-fb80-4b03-9288-ed17a0d05140", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-122910456-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "547d9a3099b94ff488ef6929ae591bac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1c797172-a569-458e-aeb0-3f21e589a740", "external-id": "nsx-vlan-transportzone-957", "segmentation_id": 957, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43370cd4-af", "ovs_interfaceid": "43370cd4-afaf-437d-8079-cf9ad90682aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1025.528540] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9b:4c:5d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1c797172-a569-458e-aeb0-3f21e589a740', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '43370cd4-afaf-437d-8079-cf9ad90682aa', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1025.536301] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Creating folder: Project (547d9a3099b94ff488ef6929ae591bac). Parent ref: group-v277447. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1025.536603] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3ca9fb7e-430a-42ea-9e7b-c55bb54b55ee {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.548453] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Created folder: Project (547d9a3099b94ff488ef6929ae591bac) in parent group-v277447. [ 1025.548692] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Creating folder: Instances. Parent ref: group-v277585. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1025.548995] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-649ae359-b190-4210-b039-7d5d582934b4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.559244] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Created folder: Instances in parent group-v277585. [ 1025.559244] env[63021]: DEBUG oslo.service.loopingcall [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1025.559462] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1025.559703] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7660b21a-d36f-46fe-8cf0-5838c584a5d6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.578748] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1025.578748] env[63021]: value = "task-1294206" [ 1025.578748] env[63021]: _type = "Task" [ 1025.578748] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.586835] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294206, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.677462] env[63021]: DEBUG oslo_concurrency.lockutils [req-d1dc68e0-7773-4ae0-b0a5-dd3faa509ea2 req-8564d181-1b0f-4113-9d2a-fbad7a2504d3 service nova] Releasing lock "refresh_cache-8d9387b8-6a84-4ada-8ac7-4ffd922b675a" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.712900] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.213s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.713543] env[63021]: DEBUG nova.compute.manager [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1025.716567] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.872s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.716799] env[63021]: DEBUG nova.objects.instance [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lazy-loading 'resources' on Instance uuid 6155c098-e4a6-47e6-b343-4a77ca90eb2e {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1025.745284] env[63021]: DEBUG nova.compute.manager [req-68e7a6d5-51f9-4552-9f9f-ccad676213c3 req-e0938e89-677c-426b-9c34-86c97596d819 service nova] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Received event network-changed-43370cd4-afaf-437d-8079-cf9ad90682aa {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1025.745492] env[63021]: DEBUG nova.compute.manager [req-68e7a6d5-51f9-4552-9f9f-ccad676213c3 req-e0938e89-677c-426b-9c34-86c97596d819 service nova] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Refreshing instance network info cache due to event network-changed-43370cd4-afaf-437d-8079-cf9ad90682aa. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1025.745809] env[63021]: DEBUG oslo_concurrency.lockutils [req-68e7a6d5-51f9-4552-9f9f-ccad676213c3 req-e0938e89-677c-426b-9c34-86c97596d819 service nova] Acquiring lock "refresh_cache-d03c47da-7e73-4474-ad9b-07d27629dcae" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.745879] env[63021]: DEBUG oslo_concurrency.lockutils [req-68e7a6d5-51f9-4552-9f9f-ccad676213c3 req-e0938e89-677c-426b-9c34-86c97596d819 service nova] Acquired lock "refresh_cache-d03c47da-7e73-4474-ad9b-07d27629dcae" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.746036] env[63021]: DEBUG nova.network.neutron [req-68e7a6d5-51f9-4552-9f9f-ccad676213c3 req-e0938e89-677c-426b-9c34-86c97596d819 service nova] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Refreshing network info cache for port 43370cd4-afaf-437d-8079-cf9ad90682aa {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1025.831961] env[63021]: DEBUG oslo_concurrency.lockutils [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.914222] env[63021]: DEBUG oslo_vmware.api [None req-b3b3cff2-7134-42ec-89e0-a3a1192740e5 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294203, 'name': ReconfigVM_Task, 'duration_secs': 0.356356} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.914573] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3b3cff2-7134-42ec-89e0-a3a1192740e5 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Reconfigured VM instance instance-0000005b to attach disk [datastore2] volume-236a8786-fa3f-411d-a486-ddd5f2950147/volume-236a8786-fa3f-411d-a486-ddd5f2950147.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1025.919482] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-416f5b88-8697-467b-b1a4-64e9b24870bc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.934120] env[63021]: DEBUG oslo_vmware.api [None req-b3b3cff2-7134-42ec-89e0-a3a1192740e5 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 1025.934120] env[63021]: value = "task-1294207" [ 1025.934120] env[63021]: _type = "Task" [ 1025.934120] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.944590] env[63021]: DEBUG oslo_vmware.api [None req-b3b3cff2-7134-42ec-89e0-a3a1192740e5 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294207, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.089805] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294206, 'name': CreateVM_Task, 'duration_secs': 0.400621} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.090015] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1026.090800] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.090985] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.091381] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1026.091686] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c0b581db-b735-406c-8f57-7271f07e9240 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.095906] env[63021]: DEBUG oslo_vmware.api [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1026.095906] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52c6c0cd-fcbd-4fd8-2cd7-a71dcd7ad74d" [ 1026.095906] env[63021]: _type = "Task" [ 1026.095906] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.103112] env[63021]: DEBUG oslo_vmware.api [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52c6c0cd-fcbd-4fd8-2cd7-a71dcd7ad74d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.220127] env[63021]: DEBUG nova.compute.utils [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1026.221621] env[63021]: DEBUG nova.compute.manager [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1026.221793] env[63021]: DEBUG nova.network.neutron [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1026.265716] env[63021]: DEBUG nova.policy [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ba16721213dd49a3b4e4990d7d319d1f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3c8959d2f1fe4706b7407c11e128b773', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 1026.372631] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3bf37f0-deea-4bce-9e83-e90e3e1c1cd4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.380483] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2669f3a-add5-4487-a023-c35317e2288c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.411836] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47e164ef-ec19-49ad-bf74-895c21360e4d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.419177] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40b4b458-2a91-4b43-9626-fa4f0ee5ff54 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.431798] env[63021]: DEBUG nova.compute.provider_tree [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1026.442767] env[63021]: DEBUG oslo_vmware.api [None req-b3b3cff2-7134-42ec-89e0-a3a1192740e5 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294207, 'name': ReconfigVM_Task, 'duration_secs': 0.183446} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.444973] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3b3cff2-7134-42ec-89e0-a3a1192740e5 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277584', 'volume_id': '236a8786-fa3f-411d-a486-ddd5f2950147', 'name': 'volume-236a8786-fa3f-411d-a486-ddd5f2950147', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '67f736d9-82ec-448e-8b72-84fdd704aa8d', 'attached_at': '', 'detached_at': '', 'volume_id': '236a8786-fa3f-411d-a486-ddd5f2950147', 'serial': '236a8786-fa3f-411d-a486-ddd5f2950147'} {{(pid=63021) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1026.478674] env[63021]: DEBUG nova.network.neutron [req-68e7a6d5-51f9-4552-9f9f-ccad676213c3 req-e0938e89-677c-426b-9c34-86c97596d819 service nova] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Updated VIF entry in instance network info cache for port 43370cd4-afaf-437d-8079-cf9ad90682aa. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1026.478717] env[63021]: DEBUG nova.network.neutron [req-68e7a6d5-51f9-4552-9f9f-ccad676213c3 req-e0938e89-677c-426b-9c34-86c97596d819 service nova] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Updating instance_info_cache with network_info: [{"id": "43370cd4-afaf-437d-8079-cf9ad90682aa", "address": "fa:16:3e:9b:4c:5d", "network": {"id": "d27bb3cc-fb80-4b03-9288-ed17a0d05140", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-122910456-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "547d9a3099b94ff488ef6929ae591bac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1c797172-a569-458e-aeb0-3f21e589a740", "external-id": "nsx-vlan-transportzone-957", "segmentation_id": 957, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43370cd4-af", "ovs_interfaceid": "43370cd4-afaf-437d-8079-cf9ad90682aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.511861] env[63021]: DEBUG nova.network.neutron [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Successfully created port: 952b66c6-27ca-4614-8875-8a94e770495a {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1026.606772] env[63021]: DEBUG oslo_vmware.api [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52c6c0cd-fcbd-4fd8-2cd7-a71dcd7ad74d, 'name': SearchDatastore_Task, 'duration_secs': 0.033957} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.607157] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.607394] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1026.607651] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.607930] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.608184] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1026.608450] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-99657929-9a95-4515-bd29-e933cb75e1d5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.616174] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1026.616361] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1026.617099] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4e1bf1e-449c-403a-8f05-50faa7d4e8bb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.622214] env[63021]: DEBUG oslo_vmware.api [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1026.622214] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52004dfd-328f-42ca-93cc-2cddd5d0e869" [ 1026.622214] env[63021]: _type = "Task" [ 1026.622214] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.629492] env[63021]: DEBUG oslo_vmware.api [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52004dfd-328f-42ca-93cc-2cddd5d0e869, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.727411] env[63021]: DEBUG nova.compute.manager [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1026.934913] env[63021]: DEBUG nova.scheduler.client.report [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1026.981728] env[63021]: DEBUG oslo_concurrency.lockutils [req-68e7a6d5-51f9-4552-9f9f-ccad676213c3 req-e0938e89-677c-426b-9c34-86c97596d819 service nova] Releasing lock "refresh_cache-d03c47da-7e73-4474-ad9b-07d27629dcae" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.133144] env[63021]: DEBUG oslo_vmware.api [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52004dfd-328f-42ca-93cc-2cddd5d0e869, 'name': SearchDatastore_Task, 'duration_secs': 0.008074} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.133946] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1eeeca83-2dd3-4391-bb34-d461b8c1d2f4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.139269] env[63021]: DEBUG oslo_vmware.api [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1027.139269] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5244fcd8-c56d-8ba1-4c7e-fec6d147f727" [ 1027.139269] env[63021]: _type = "Task" [ 1027.139269] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.146822] env[63021]: DEBUG oslo_vmware.api [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5244fcd8-c56d-8ba1-4c7e-fec6d147f727, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.439878] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.723s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.443526] env[63021]: DEBUG oslo_concurrency.lockutils [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.545s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.443803] env[63021]: DEBUG nova.objects.instance [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lazy-loading 'resources' on Instance uuid 67aed8eb-d08c-47cc-8558-5c5a0f058d6e {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1027.469064] env[63021]: INFO nova.scheduler.client.report [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Deleted allocations for instance 6155c098-e4a6-47e6-b343-4a77ca90eb2e [ 1027.479643] env[63021]: DEBUG nova.objects.instance [None req-b3b3cff2-7134-42ec-89e0-a3a1192740e5 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lazy-loading 'flavor' on Instance uuid 67f736d9-82ec-448e-8b72-84fdd704aa8d {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1027.650854] env[63021]: DEBUG oslo_vmware.api [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5244fcd8-c56d-8ba1-4c7e-fec6d147f727, 'name': SearchDatastore_Task, 'duration_secs': 0.008722} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.651167] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.651453] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] d03c47da-7e73-4474-ad9b-07d27629dcae/d03c47da-7e73-4474-ad9b-07d27629dcae.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1027.651718] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-55445796-1de2-4f91-af8b-8a47f94a73bb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.658279] env[63021]: DEBUG oslo_vmware.api [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1027.658279] env[63021]: value = "task-1294208" [ 1027.658279] env[63021]: _type = "Task" [ 1027.658279] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.669589] env[63021]: DEBUG oslo_vmware.api [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294208, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.738430] env[63021]: DEBUG nova.compute.manager [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1027.769179] env[63021]: DEBUG nova.virt.hardware [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1027.769504] env[63021]: DEBUG nova.virt.hardware [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1027.769697] env[63021]: DEBUG nova.virt.hardware [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1027.769919] env[63021]: DEBUG nova.virt.hardware [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1027.770127] env[63021]: DEBUG nova.virt.hardware [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1027.770312] env[63021]: DEBUG nova.virt.hardware [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1027.770568] env[63021]: DEBUG nova.virt.hardware [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1027.770760] env[63021]: DEBUG nova.virt.hardware [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1027.770958] env[63021]: DEBUG nova.virt.hardware [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1027.771172] env[63021]: DEBUG nova.virt.hardware [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1027.771374] env[63021]: DEBUG nova.virt.hardware [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1027.772538] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d115186-d230-41db-9d26-47b5be5189a6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.781533] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09bf1d10-8ef0-495d-91b4-5f6b80db4dd2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.890707] env[63021]: DEBUG nova.compute.manager [req-54d02aa6-57e8-4f74-9993-2d297a0632f7 req-4bdd7fb8-fb6f-41a9-a009-091b7cbf9f32 service nova] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Received event network-vif-plugged-952b66c6-27ca-4614-8875-8a94e770495a {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1027.890955] env[63021]: DEBUG oslo_concurrency.lockutils [req-54d02aa6-57e8-4f74-9993-2d297a0632f7 req-4bdd7fb8-fb6f-41a9-a009-091b7cbf9f32 service nova] Acquiring lock "63d5b0e4-0d61-4fa7-b193-2eaa171fc461-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.891207] env[63021]: DEBUG oslo_concurrency.lockutils [req-54d02aa6-57e8-4f74-9993-2d297a0632f7 req-4bdd7fb8-fb6f-41a9-a009-091b7cbf9f32 service nova] Lock "63d5b0e4-0d61-4fa7-b193-2eaa171fc461-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.891378] env[63021]: DEBUG oslo_concurrency.lockutils [req-54d02aa6-57e8-4f74-9993-2d297a0632f7 req-4bdd7fb8-fb6f-41a9-a009-091b7cbf9f32 service nova] Lock "63d5b0e4-0d61-4fa7-b193-2eaa171fc461-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.891574] env[63021]: DEBUG nova.compute.manager [req-54d02aa6-57e8-4f74-9993-2d297a0632f7 req-4bdd7fb8-fb6f-41a9-a009-091b7cbf9f32 service nova] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] No waiting events found dispatching network-vif-plugged-952b66c6-27ca-4614-8875-8a94e770495a {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1027.891704] env[63021]: WARNING nova.compute.manager [req-54d02aa6-57e8-4f74-9993-2d297a0632f7 req-4bdd7fb8-fb6f-41a9-a009-091b7cbf9f32 service nova] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Received unexpected event network-vif-plugged-952b66c6-27ca-4614-8875-8a94e770495a for instance with vm_state building and task_state spawning. [ 1027.976867] env[63021]: DEBUG oslo_concurrency.lockutils [None req-4797e2ff-0df3-41da-ae33-bac12c45b9c5 tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "6155c098-e4a6-47e6-b343-4a77ca90eb2e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.554s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.986493] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b3b3cff2-7134-42ec-89e0-a3a1192740e5 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "67f736d9-82ec-448e-8b72-84fdd704aa8d" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.246s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.113719] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7f7baa5-f9e6-4f17-b682-a42630e597a5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.121580] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09d80c6f-f098-464c-a055-87ea6f20fd78 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.153172] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bc99db4-2649-409d-a70e-1eaabdefd64f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.163105] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5699f41-3746-4de2-9717-1aac4d7f99ec {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.166791] env[63021]: INFO nova.compute.manager [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Rescuing [ 1028.167040] env[63021]: DEBUG oslo_concurrency.lockutils [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "refresh_cache-67f736d9-82ec-448e-8b72-84fdd704aa8d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.167210] env[63021]: DEBUG oslo_concurrency.lockutils [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquired lock "refresh_cache-67f736d9-82ec-448e-8b72-84fdd704aa8d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.167374] env[63021]: DEBUG nova.network.neutron [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1028.174206] env[63021]: DEBUG oslo_vmware.api [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294208, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.461069} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.182606] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] d03c47da-7e73-4474-ad9b-07d27629dcae/d03c47da-7e73-4474-ad9b-07d27629dcae.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1028.182883] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1028.183635] env[63021]: DEBUG nova.compute.provider_tree [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1028.185133] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-07e99d07-3f05-4373-8744-1c8faca4ff45 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.192288] env[63021]: DEBUG oslo_vmware.api [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1028.192288] env[63021]: value = "task-1294209" [ 1028.192288] env[63021]: _type = "Task" [ 1028.192288] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.200471] env[63021]: DEBUG oslo_vmware.api [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294209, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.244338] env[63021]: DEBUG oslo_concurrency.lockutils [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "2d647028-306a-41d5-96f0-2e82a98ab56d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.244621] env[63021]: DEBUG oslo_concurrency.lockutils [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "2d647028-306a-41d5-96f0-2e82a98ab56d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.244837] env[63021]: DEBUG oslo_concurrency.lockutils [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "2d647028-306a-41d5-96f0-2e82a98ab56d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.245070] env[63021]: DEBUG oslo_concurrency.lockutils [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "2d647028-306a-41d5-96f0-2e82a98ab56d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.245263] env[63021]: DEBUG oslo_concurrency.lockutils [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "2d647028-306a-41d5-96f0-2e82a98ab56d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.247551] env[63021]: INFO nova.compute.manager [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Terminating instance [ 1028.249579] env[63021]: DEBUG nova.compute.manager [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1028.249768] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1028.250611] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-894d2fab-c1e6-482a-ac5e-b514bb7a497c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.258553] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1028.258800] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b713cf88-f656-436f-a3d9-9d31edb17f67 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.264980] env[63021]: DEBUG oslo_vmware.api [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 1028.264980] env[63021]: value = "task-1294210" [ 1028.264980] env[63021]: _type = "Task" [ 1028.264980] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.272639] env[63021]: DEBUG oslo_vmware.api [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294210, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.454817] env[63021]: DEBUG nova.network.neutron [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Successfully updated port: 952b66c6-27ca-4614-8875-8a94e770495a {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1028.475754] env[63021]: DEBUG nova.compute.manager [req-75c21bfa-cd5c-469a-a85c-78e7c25e9baa req-facb3a31-f21f-43d0-bae2-54a0bffd402b service nova] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Received event network-changed-952b66c6-27ca-4614-8875-8a94e770495a {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1028.475963] env[63021]: DEBUG nova.compute.manager [req-75c21bfa-cd5c-469a-a85c-78e7c25e9baa req-facb3a31-f21f-43d0-bae2-54a0bffd402b service nova] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Refreshing instance network info cache due to event network-changed-952b66c6-27ca-4614-8875-8a94e770495a. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1028.476200] env[63021]: DEBUG oslo_concurrency.lockutils [req-75c21bfa-cd5c-469a-a85c-78e7c25e9baa req-facb3a31-f21f-43d0-bae2-54a0bffd402b service nova] Acquiring lock "refresh_cache-63d5b0e4-0d61-4fa7-b193-2eaa171fc461" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.476348] env[63021]: DEBUG oslo_concurrency.lockutils [req-75c21bfa-cd5c-469a-a85c-78e7c25e9baa req-facb3a31-f21f-43d0-bae2-54a0bffd402b service nova] Acquired lock "refresh_cache-63d5b0e4-0d61-4fa7-b193-2eaa171fc461" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.476515] env[63021]: DEBUG nova.network.neutron [req-75c21bfa-cd5c-469a-a85c-78e7c25e9baa req-facb3a31-f21f-43d0-bae2-54a0bffd402b service nova] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Refreshing network info cache for port 952b66c6-27ca-4614-8875-8a94e770495a {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1028.686314] env[63021]: DEBUG nova.scheduler.client.report [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1028.701349] env[63021]: DEBUG oslo_vmware.api [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294209, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070389} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.704250] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1028.705041] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f22f850e-6a60-4ff7-8da3-d1c36569c354 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.727474] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] d03c47da-7e73-4474-ad9b-07d27629dcae/d03c47da-7e73-4474-ad9b-07d27629dcae.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1028.727980] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5cfdaf7f-00b4-4024-91a1-eee905678685 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.749384] env[63021]: DEBUG oslo_vmware.api [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1028.749384] env[63021]: value = "task-1294211" [ 1028.749384] env[63021]: _type = "Task" [ 1028.749384] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.758412] env[63021]: DEBUG oslo_vmware.api [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294211, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.773189] env[63021]: DEBUG oslo_vmware.api [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294210, 'name': PowerOffVM_Task, 'duration_secs': 0.182174} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.773466] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1028.773654] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1028.773926] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4ec1aa09-d524-4da1-bf1e-a1531dfdcbc0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.844080] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1028.844317] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1028.844514] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Deleting the datastore file [datastore1] 2d647028-306a-41d5-96f0-2e82a98ab56d {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1028.844787] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-66a4d795-9583-46ec-89b1-bad6339620d0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.851498] env[63021]: DEBUG oslo_vmware.api [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for the task: (returnval){ [ 1028.851498] env[63021]: value = "task-1294213" [ 1028.851498] env[63021]: _type = "Task" [ 1028.851498] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.860877] env[63021]: DEBUG oslo_vmware.api [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294213, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.894511] env[63021]: DEBUG nova.network.neutron [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Updating instance_info_cache with network_info: [{"id": "db333f88-b16f-439c-99fe-ca5884a09dc1", "address": "fa:16:3e:40:f8:6d", "network": {"id": "3aaeed35-f452-4144-a259-8f440f8a8bf4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-229941363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.192", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b672b1d7d7ad454ebfa3e2fd2421c754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb333f88-b1", "ovs_interfaceid": "db333f88-b16f-439c-99fe-ca5884a09dc1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.957061] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "refresh_cache-63d5b0e4-0d61-4fa7-b193-2eaa171fc461" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.009102] env[63021]: DEBUG nova.network.neutron [req-75c21bfa-cd5c-469a-a85c-78e7c25e9baa req-facb3a31-f21f-43d0-bae2-54a0bffd402b service nova] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1029.079413] env[63021]: DEBUG nova.network.neutron [req-75c21bfa-cd5c-469a-a85c-78e7c25e9baa req-facb3a31-f21f-43d0-bae2-54a0bffd402b service nova] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.190887] env[63021]: DEBUG oslo_concurrency.lockutils [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.747s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.193327] env[63021]: DEBUG oslo_concurrency.lockutils [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.361s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.195199] env[63021]: INFO nova.compute.claims [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1029.209618] env[63021]: INFO nova.scheduler.client.report [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Deleted allocations for instance 67aed8eb-d08c-47cc-8558-5c5a0f058d6e [ 1029.259957] env[63021]: DEBUG oslo_vmware.api [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294211, 'name': ReconfigVM_Task, 'duration_secs': 0.265839} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.260344] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Reconfigured VM instance instance-00000062 to attach disk [datastore2] d03c47da-7e73-4474-ad9b-07d27629dcae/d03c47da-7e73-4474-ad9b-07d27629dcae.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1029.260986] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4c80be6b-1509-4e0f-b103-fb9a288fe177 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.267140] env[63021]: DEBUG oslo_vmware.api [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1029.267140] env[63021]: value = "task-1294214" [ 1029.267140] env[63021]: _type = "Task" [ 1029.267140] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.275692] env[63021]: DEBUG oslo_vmware.api [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294214, 'name': Rename_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.361619] env[63021]: DEBUG oslo_vmware.api [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Task: {'id': task-1294213, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.204265} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.361916] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1029.362295] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1029.362431] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1029.362494] env[63021]: INFO nova.compute.manager [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1029.362741] env[63021]: DEBUG oslo.service.loopingcall [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1029.362979] env[63021]: DEBUG nova.compute.manager [-] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1029.363076] env[63021]: DEBUG nova.network.neutron [-] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1029.397615] env[63021]: DEBUG oslo_concurrency.lockutils [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Releasing lock "refresh_cache-67f736d9-82ec-448e-8b72-84fdd704aa8d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.582022] env[63021]: DEBUG oslo_concurrency.lockutils [req-75c21bfa-cd5c-469a-a85c-78e7c25e9baa req-facb3a31-f21f-43d0-bae2-54a0bffd402b service nova] Releasing lock "refresh_cache-63d5b0e4-0d61-4fa7-b193-2eaa171fc461" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.582266] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquired lock "refresh_cache-63d5b0e4-0d61-4fa7-b193-2eaa171fc461" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.582576] env[63021]: DEBUG nova.network.neutron [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1029.715950] env[63021]: DEBUG oslo_concurrency.lockutils [None req-63c441f9-2b0d-4425-9292-6e2c3bb2d0f8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "67aed8eb-d08c-47cc-8558-5c5a0f058d6e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.834s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.777437] env[63021]: DEBUG oslo_vmware.api [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294214, 'name': Rename_Task, 'duration_secs': 0.14834} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.777776] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1029.778060] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6a05b0e7-1e25-4ec5-b0fd-01e2a4986e2f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.784260] env[63021]: DEBUG oslo_vmware.api [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1029.784260] env[63021]: value = "task-1294215" [ 1029.784260] env[63021]: _type = "Task" [ 1029.784260] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.792326] env[63021]: DEBUG oslo_vmware.api [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294215, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.929478] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1029.929756] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f07fc15b-cc8f-4ade-aa4c-6ee9622da7d5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.936017] env[63021]: DEBUG oslo_vmware.api [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 1029.936017] env[63021]: value = "task-1294216" [ 1029.936017] env[63021]: _type = "Task" [ 1029.936017] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.945383] env[63021]: DEBUG oslo_vmware.api [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294216, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.094138] env[63021]: DEBUG nova.network.neutron [-] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.136162] env[63021]: DEBUG nova.network.neutron [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1030.298038] env[63021]: DEBUG oslo_vmware.api [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294215, 'name': PowerOnVM_Task, 'duration_secs': 0.483538} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.298617] env[63021]: DEBUG nova.network.neutron [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Updating instance_info_cache with network_info: [{"id": "952b66c6-27ca-4614-8875-8a94e770495a", "address": "fa:16:3e:dc:95:2a", "network": {"id": "0c0dd1a7-6687-45c5-82b1-33f34bdbb848", "bridge": "br-int", "label": "tempest-ServersTestJSON-523240467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c8959d2f1fe4706b7407c11e128b773", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffc811e4-0e4e-4748-8c7e-b3f14ccbd42d", "external-id": "nsx-vlan-transportzone-404", "segmentation_id": 404, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap952b66c6-27", "ovs_interfaceid": "952b66c6-27ca-4614-8875-8a94e770495a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.300210] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1030.300434] env[63021]: INFO nova.compute.manager [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Took 7.05 seconds to spawn the instance on the hypervisor. [ 1030.300612] env[63021]: DEBUG nova.compute.manager [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1030.301620] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b390df90-aa1a-45f6-a5b6-f002f78ca580 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.359855] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13838ae0-0668-4ed8-840a-488cfd84b623 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.367929] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26e0365a-eb3d-439f-b60a-29d90faae2d2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.398352] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c8274a9-8eb6-4a4e-a65a-5ff02cf47381 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.405853] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1c5b2cd-a908-485f-bc3e-7528efed8452 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.421960] env[63021]: DEBUG nova.compute.provider_tree [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1030.446105] env[63021]: DEBUG oslo_vmware.api [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294216, 'name': PowerOffVM_Task, 'duration_secs': 0.252786} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.446366] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1030.447163] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3db5189-a551-414c-8404-b38efc5d4ddd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.470028] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cae530b-1b7e-4190-995b-b2f5a2ae3c43 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.502264] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1030.502899] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-95539a96-0cb9-4ee7-b672-948d046533be {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.506779] env[63021]: DEBUG nova.compute.manager [req-cc3d5b79-ec2c-426a-b704-8bdb3b366f7a req-5d261c4b-ad1c-441f-a326-1c456408cd78 service nova] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Received event network-vif-deleted-eab7f4fe-4832-40e4-aed2-0e931d91a4d7 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1030.512100] env[63021]: DEBUG oslo_vmware.api [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 1030.512100] env[63021]: value = "task-1294217" [ 1030.512100] env[63021]: _type = "Task" [ 1030.512100] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.522270] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] VM already powered off {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1030.522270] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1030.522534] env[63021]: DEBUG oslo_concurrency.lockutils [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.522724] env[63021]: DEBUG oslo_concurrency.lockutils [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.522837] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1030.523098] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-823e86b3-93fa-4cc9-90ea-cb635b721d5b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.538118] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1030.538118] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1030.538877] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-90d058fa-f599-49d5-bd4e-6ce66a186e42 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.544440] env[63021]: DEBUG oslo_vmware.api [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 1030.544440] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52446e87-d6a9-ddc2-4dd8-23e9de16dbac" [ 1030.544440] env[63021]: _type = "Task" [ 1030.544440] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.551937] env[63021]: DEBUG oslo_vmware.api [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52446e87-d6a9-ddc2-4dd8-23e9de16dbac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.596983] env[63021]: INFO nova.compute.manager [-] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Took 1.23 seconds to deallocate network for instance. [ 1030.801743] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Releasing lock "refresh_cache-63d5b0e4-0d61-4fa7-b193-2eaa171fc461" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.802118] env[63021]: DEBUG nova.compute.manager [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Instance network_info: |[{"id": "952b66c6-27ca-4614-8875-8a94e770495a", "address": "fa:16:3e:dc:95:2a", "network": {"id": "0c0dd1a7-6687-45c5-82b1-33f34bdbb848", "bridge": "br-int", "label": "tempest-ServersTestJSON-523240467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c8959d2f1fe4706b7407c11e128b773", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffc811e4-0e4e-4748-8c7e-b3f14ccbd42d", "external-id": "nsx-vlan-transportzone-404", "segmentation_id": 404, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap952b66c6-27", "ovs_interfaceid": "952b66c6-27ca-4614-8875-8a94e770495a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1030.802608] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dc:95:2a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ffc811e4-0e4e-4748-8c7e-b3f14ccbd42d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '952b66c6-27ca-4614-8875-8a94e770495a', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1030.810687] env[63021]: DEBUG oslo.service.loopingcall [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1030.810924] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1030.811170] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ce111176-6135-4e26-ad53-e41b5b5804c2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.836729] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1030.836729] env[63021]: value = "task-1294218" [ 1030.836729] env[63021]: _type = "Task" [ 1030.836729] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.837228] env[63021]: INFO nova.compute.manager [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Took 13.55 seconds to build instance. [ 1030.847189] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294218, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.924989] env[63021]: DEBUG nova.scheduler.client.report [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1031.054711] env[63021]: DEBUG oslo_vmware.api [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52446e87-d6a9-ddc2-4dd8-23e9de16dbac, 'name': SearchDatastore_Task, 'duration_secs': 0.041958} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.055660] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f6716e1-7939-4768-a7e1-0259d17cfdab {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.061576] env[63021]: DEBUG oslo_vmware.api [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 1031.061576] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]520e3293-ed95-8941-e170-16cc23121c84" [ 1031.061576] env[63021]: _type = "Task" [ 1031.061576] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.070770] env[63021]: DEBUG oslo_vmware.api [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]520e3293-ed95-8941-e170-16cc23121c84, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.103629] env[63021]: DEBUG oslo_concurrency.lockutils [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.312916] env[63021]: DEBUG nova.compute.manager [req-1edc0397-41f8-40e3-ae06-6e1f2c26b6fa req-8bc19b18-c508-46d2-91b5-0bb3ee3d5e51 service nova] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Received event network-changed-43370cd4-afaf-437d-8079-cf9ad90682aa {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1031.313122] env[63021]: DEBUG nova.compute.manager [req-1edc0397-41f8-40e3-ae06-6e1f2c26b6fa req-8bc19b18-c508-46d2-91b5-0bb3ee3d5e51 service nova] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Refreshing instance network info cache due to event network-changed-43370cd4-afaf-437d-8079-cf9ad90682aa. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1031.313312] env[63021]: DEBUG oslo_concurrency.lockutils [req-1edc0397-41f8-40e3-ae06-6e1f2c26b6fa req-8bc19b18-c508-46d2-91b5-0bb3ee3d5e51 service nova] Acquiring lock "refresh_cache-d03c47da-7e73-4474-ad9b-07d27629dcae" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.313486] env[63021]: DEBUG oslo_concurrency.lockutils [req-1edc0397-41f8-40e3-ae06-6e1f2c26b6fa req-8bc19b18-c508-46d2-91b5-0bb3ee3d5e51 service nova] Acquired lock "refresh_cache-d03c47da-7e73-4474-ad9b-07d27629dcae" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.313806] env[63021]: DEBUG nova.network.neutron [req-1edc0397-41f8-40e3-ae06-6e1f2c26b6fa req-8bc19b18-c508-46d2-91b5-0bb3ee3d5e51 service nova] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Refreshing network info cache for port 43370cd4-afaf-437d-8079-cf9ad90682aa {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1031.343248] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fdc4492b-146b-440c-ae86-a74dff3b0dd1 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "d03c47da-7e73-4474-ad9b-07d27629dcae" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.063s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.349750] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294218, 'name': CreateVM_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.431376] env[63021]: DEBUG oslo_concurrency.lockutils [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.238s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.431913] env[63021]: DEBUG nova.compute.manager [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1031.434984] env[63021]: DEBUG oslo_concurrency.lockutils [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.331s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.435224] env[63021]: DEBUG nova.objects.instance [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lazy-loading 'resources' on Instance uuid 2d647028-306a-41d5-96f0-2e82a98ab56d {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1031.571857] env[63021]: DEBUG oslo_vmware.api [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]520e3293-ed95-8941-e170-16cc23121c84, 'name': SearchDatastore_Task, 'duration_secs': 0.014106} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.572154] env[63021]: DEBUG oslo_concurrency.lockutils [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.572415] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 67f736d9-82ec-448e-8b72-84fdd704aa8d/e624225d-af0f-415c-90cc-d4b228c479f6-rescue.vmdk. {{(pid=63021) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1031.572914] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6e0bec8e-2ec6-456b-8db2-e83ae4793fdf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.579716] env[63021]: DEBUG oslo_vmware.api [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 1031.579716] env[63021]: value = "task-1294219" [ 1031.579716] env[63021]: _type = "Task" [ 1031.579716] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.587326] env[63021]: DEBUG oslo_vmware.api [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294219, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.851482] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294218, 'name': CreateVM_Task, 'duration_secs': 0.596066} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.854285] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1031.855039] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.855238] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.855619] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1031.856367] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-420504a9-30ab-4abd-b6f5-4ab1d309e241 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.861829] env[63021]: DEBUG oslo_vmware.api [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1031.861829] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52e0bf61-b474-7a01-b3f3-6b263c36aca2" [ 1031.861829] env[63021]: _type = "Task" [ 1031.861829] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.871802] env[63021]: DEBUG oslo_vmware.api [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52e0bf61-b474-7a01-b3f3-6b263c36aca2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.938742] env[63021]: DEBUG nova.compute.utils [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1031.943689] env[63021]: DEBUG nova.compute.manager [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1031.944060] env[63021]: DEBUG nova.network.neutron [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1031.989448] env[63021]: DEBUG nova.policy [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b427f7d1317341f38754a2b324f4f5aa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e238d5c1668540d2a9fffd3fd832d9b8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 1032.032924] env[63021]: DEBUG nova.compute.manager [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Stashing vm_state: active {{(pid=63021) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1032.093616] env[63021]: DEBUG oslo_vmware.api [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294219, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.094915] env[63021]: DEBUG nova.network.neutron [req-1edc0397-41f8-40e3-ae06-6e1f2c26b6fa req-8bc19b18-c508-46d2-91b5-0bb3ee3d5e51 service nova] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Updated VIF entry in instance network info cache for port 43370cd4-afaf-437d-8079-cf9ad90682aa. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1032.095683] env[63021]: DEBUG nova.network.neutron [req-1edc0397-41f8-40e3-ae06-6e1f2c26b6fa req-8bc19b18-c508-46d2-91b5-0bb3ee3d5e51 service nova] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Updating instance_info_cache with network_info: [{"id": "43370cd4-afaf-437d-8079-cf9ad90682aa", "address": "fa:16:3e:9b:4c:5d", "network": {"id": "d27bb3cc-fb80-4b03-9288-ed17a0d05140", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-122910456-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.163", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "547d9a3099b94ff488ef6929ae591bac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1c797172-a569-458e-aeb0-3f21e589a740", "external-id": "nsx-vlan-transportzone-957", "segmentation_id": 957, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43370cd4-af", "ovs_interfaceid": "43370cd4-afaf-437d-8079-cf9ad90682aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.135595] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b43596bf-d8a1-4e17-acf2-4a4b7825b55d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.147431] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dee82bc2-de6c-4ded-a694-74a21d1d45b5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.183936] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c440e02-31ed-456a-95cf-568d101ee800 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.194525] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eafebb0-e192-479e-a7b2-ff52595a92aa {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.210316] env[63021]: DEBUG nova.compute.provider_tree [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1032.373088] env[63021]: DEBUG oslo_vmware.api [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52e0bf61-b474-7a01-b3f3-6b263c36aca2, 'name': SearchDatastore_Task, 'duration_secs': 0.014744} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.373421] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.373659] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1032.373892] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.374124] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.374239] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1032.374510] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-20656579-b105-4e3f-90b2-528e2829b108 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.382119] env[63021]: DEBUG nova.network.neutron [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Successfully created port: 1888ca08-3a84-45c4-9fe2-79295174318d {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1032.385232] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1032.385420] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1032.386165] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-81a98e63-e2c5-45b6-812f-d75a6a7b3582 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.391378] env[63021]: DEBUG oslo_vmware.api [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1032.391378] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5262bf89-79b3-6958-ecdf-95d5d9820042" [ 1032.391378] env[63021]: _type = "Task" [ 1032.391378] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.402511] env[63021]: DEBUG oslo_vmware.api [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5262bf89-79b3-6958-ecdf-95d5d9820042, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.444480] env[63021]: DEBUG nova.compute.manager [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1032.550856] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.590149] env[63021]: DEBUG oslo_vmware.api [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294219, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.692283} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.590438] env[63021]: INFO nova.virt.vmwareapi.ds_util [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 67f736d9-82ec-448e-8b72-84fdd704aa8d/e624225d-af0f-415c-90cc-d4b228c479f6-rescue.vmdk. [ 1032.591209] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5745608-ddcc-4000-b1f9-d03f79ea85f1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.611162] env[63021]: DEBUG oslo_concurrency.lockutils [req-1edc0397-41f8-40e3-ae06-6e1f2c26b6fa req-8bc19b18-c508-46d2-91b5-0bb3ee3d5e51 service nova] Releasing lock "refresh_cache-d03c47da-7e73-4474-ad9b-07d27629dcae" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.619326] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] 67f736d9-82ec-448e-8b72-84fdd704aa8d/e624225d-af0f-415c-90cc-d4b228c479f6-rescue.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1032.619617] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eebd162c-3e4e-462b-8b24-b4f89e2fa170 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.637861] env[63021]: DEBUG oslo_vmware.api [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 1032.637861] env[63021]: value = "task-1294220" [ 1032.637861] env[63021]: _type = "Task" [ 1032.637861] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.645475] env[63021]: DEBUG oslo_vmware.api [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294220, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.715110] env[63021]: DEBUG nova.scheduler.client.report [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1032.905658] env[63021]: DEBUG oslo_vmware.api [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5262bf89-79b3-6958-ecdf-95d5d9820042, 'name': SearchDatastore_Task, 'duration_secs': 0.010369} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.906615] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32d5fc1b-1c2d-4180-bf4e-05bf74b38fd9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.913772] env[63021]: DEBUG oslo_vmware.api [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1032.913772] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52e68c55-8f0f-b391-3777-7612f6a80a86" [ 1032.913772] env[63021]: _type = "Task" [ 1032.913772] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.920384] env[63021]: DEBUG oslo_vmware.api [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52e68c55-8f0f-b391-3777-7612f6a80a86, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.148355] env[63021]: DEBUG oslo_vmware.api [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294220, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.220858] env[63021]: DEBUG oslo_concurrency.lockutils [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.786s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.223339] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.673s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.242351] env[63021]: INFO nova.scheduler.client.report [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Deleted allocations for instance 2d647028-306a-41d5-96f0-2e82a98ab56d [ 1033.423069] env[63021]: DEBUG oslo_vmware.api [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52e68c55-8f0f-b391-3777-7612f6a80a86, 'name': SearchDatastore_Task, 'duration_secs': 0.010906} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.423383] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.423646] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 63d5b0e4-0d61-4fa7-b193-2eaa171fc461/63d5b0e4-0d61-4fa7-b193-2eaa171fc461.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1033.423917] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1e9df7a0-7dd6-4bdf-ad5a-2c0f18dc73dc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.430491] env[63021]: DEBUG oslo_vmware.api [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1033.430491] env[63021]: value = "task-1294221" [ 1033.430491] env[63021]: _type = "Task" [ 1033.430491] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.438879] env[63021]: DEBUG oslo_vmware.api [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294221, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.455099] env[63021]: DEBUG nova.compute.manager [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1033.479567] env[63021]: DEBUG nova.virt.hardware [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1033.479805] env[63021]: DEBUG nova.virt.hardware [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1033.479963] env[63021]: DEBUG nova.virt.hardware [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1033.480163] env[63021]: DEBUG nova.virt.hardware [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1033.480329] env[63021]: DEBUG nova.virt.hardware [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1033.480459] env[63021]: DEBUG nova.virt.hardware [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1033.480666] env[63021]: DEBUG nova.virt.hardware [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1033.480827] env[63021]: DEBUG nova.virt.hardware [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1033.480995] env[63021]: DEBUG nova.virt.hardware [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1033.481177] env[63021]: DEBUG nova.virt.hardware [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1033.481346] env[63021]: DEBUG nova.virt.hardware [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1033.482183] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96ccaf4e-7258-4e1f-a2e5-2e506e40c6e4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.489776] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b6b56d7-d1d8-462c-8ade-759429e93de9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.649379] env[63021]: DEBUG oslo_vmware.api [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294220, 'name': ReconfigVM_Task, 'duration_secs': 0.603149} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.649712] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Reconfigured VM instance instance-0000005b to attach disk [datastore1] 67f736d9-82ec-448e-8b72-84fdd704aa8d/e624225d-af0f-415c-90cc-d4b228c479f6-rescue.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1033.650648] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15c522bc-1467-4725-993d-de7992459c3c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.679123] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-849caf5e-53c7-477e-8e68-d033d3a5a267 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.693749] env[63021]: DEBUG oslo_vmware.api [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 1033.693749] env[63021]: value = "task-1294222" [ 1033.693749] env[63021]: _type = "Task" [ 1033.693749] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.701330] env[63021]: DEBUG oslo_vmware.api [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294222, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.728342] env[63021]: INFO nova.compute.claims [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1033.750702] env[63021]: DEBUG oslo_concurrency.lockutils [None req-afa147d8-e585-492b-b991-f7961c5703fe tempest-AttachInterfacesTestJSON-1061982398 tempest-AttachInterfacesTestJSON-1061982398-project-member] Lock "2d647028-306a-41d5-96f0-2e82a98ab56d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.506s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.766453] env[63021]: DEBUG nova.compute.manager [req-9dba3ea0-e035-4755-8b61-222191bd650f req-cdc24e80-b12b-47df-bf60-d37d5990679e service nova] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Received event network-vif-plugged-1888ca08-3a84-45c4-9fe2-79295174318d {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1033.766707] env[63021]: DEBUG oslo_concurrency.lockutils [req-9dba3ea0-e035-4755-8b61-222191bd650f req-cdc24e80-b12b-47df-bf60-d37d5990679e service nova] Acquiring lock "acc38ba0-515d-41d4-81dd-2b4374e033e7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.766921] env[63021]: DEBUG oslo_concurrency.lockutils [req-9dba3ea0-e035-4755-8b61-222191bd650f req-cdc24e80-b12b-47df-bf60-d37d5990679e service nova] Lock "acc38ba0-515d-41d4-81dd-2b4374e033e7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.767314] env[63021]: DEBUG oslo_concurrency.lockutils [req-9dba3ea0-e035-4755-8b61-222191bd650f req-cdc24e80-b12b-47df-bf60-d37d5990679e service nova] Lock "acc38ba0-515d-41d4-81dd-2b4374e033e7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.767314] env[63021]: DEBUG nova.compute.manager [req-9dba3ea0-e035-4755-8b61-222191bd650f req-cdc24e80-b12b-47df-bf60-d37d5990679e service nova] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] No waiting events found dispatching network-vif-plugged-1888ca08-3a84-45c4-9fe2-79295174318d {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1033.767468] env[63021]: WARNING nova.compute.manager [req-9dba3ea0-e035-4755-8b61-222191bd650f req-cdc24e80-b12b-47df-bf60-d37d5990679e service nova] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Received unexpected event network-vif-plugged-1888ca08-3a84-45c4-9fe2-79295174318d for instance with vm_state building and task_state spawning. [ 1033.853576] env[63021]: DEBUG nova.network.neutron [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Successfully updated port: 1888ca08-3a84-45c4-9fe2-79295174318d {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1033.943675] env[63021]: DEBUG oslo_vmware.api [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294221, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.203508] env[63021]: DEBUG oslo_vmware.api [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294222, 'name': ReconfigVM_Task, 'duration_secs': 0.168299} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.203874] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1034.204119] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b3e784eb-e628-4e34-883a-47875338faaa {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.210372] env[63021]: DEBUG oslo_vmware.api [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 1034.210372] env[63021]: value = "task-1294223" [ 1034.210372] env[63021]: _type = "Task" [ 1034.210372] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.217674] env[63021]: DEBUG oslo_vmware.api [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294223, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.235033] env[63021]: INFO nova.compute.resource_tracker [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Updating resource usage from migration 3d2ed986-6753-426d-b8c0-b7b9ad33739e [ 1034.356917] env[63021]: DEBUG oslo_concurrency.lockutils [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "refresh_cache-acc38ba0-515d-41d4-81dd-2b4374e033e7" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.356917] env[63021]: DEBUG oslo_concurrency.lockutils [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquired lock "refresh_cache-acc38ba0-515d-41d4-81dd-2b4374e033e7" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.356917] env[63021]: DEBUG nova.network.neutron [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1034.372446] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cba16cf-e17e-46f1-84d9-6668e70b3fb6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.380895] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ced17d34-d924-488b-b10b-497b579a505a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.411487] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09288018-4e34-4069-9580-b27466fb231b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.418924] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d02d6a2-224e-42fd-9256-056ebcb3552d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.432724] env[63021]: DEBUG nova.compute.provider_tree [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1034.441520] env[63021]: DEBUG oslo_vmware.api [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294221, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.810254} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.442310] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 63d5b0e4-0d61-4fa7-b193-2eaa171fc461/63d5b0e4-0d61-4fa7-b193-2eaa171fc461.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1034.442540] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1034.442792] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dd24497d-7983-4aa2-ab31-6afc0d6b2f6a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.449912] env[63021]: DEBUG oslo_vmware.api [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1034.449912] env[63021]: value = "task-1294224" [ 1034.449912] env[63021]: _type = "Task" [ 1034.449912] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.458746] env[63021]: DEBUG oslo_vmware.api [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294224, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.719745] env[63021]: DEBUG oslo_vmware.api [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294223, 'name': PowerOnVM_Task, 'duration_secs': 0.426535} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.720094] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1034.723300] env[63021]: DEBUG nova.compute.manager [None req-26f60cc5-c3be-4be9-8038-09eb7aa73b48 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1034.724107] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f34a6385-3092-46c1-a503-eb950ab30719 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.888720] env[63021]: DEBUG nova.network.neutron [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1034.937983] env[63021]: DEBUG nova.scheduler.client.report [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1034.962576] env[63021]: DEBUG oslo_vmware.api [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294224, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068361} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.962823] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1034.963613] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ed11fbf-e414-4166-a424-94ab595eb1c9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.985566] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] 63d5b0e4-0d61-4fa7-b193-2eaa171fc461/63d5b0e4-0d61-4fa7-b193-2eaa171fc461.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1034.985840] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2875af5b-2db1-4736-b627-859c30e77365 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.008238] env[63021]: DEBUG oslo_vmware.api [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1035.008238] env[63021]: value = "task-1294225" [ 1035.008238] env[63021]: _type = "Task" [ 1035.008238] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.016202] env[63021]: DEBUG oslo_vmware.api [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294225, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.035395] env[63021]: DEBUG nova.network.neutron [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Updating instance_info_cache with network_info: [{"id": "1888ca08-3a84-45c4-9fe2-79295174318d", "address": "fa:16:3e:d9:b4:fb", "network": {"id": "81917a55-4e41-4898-9f4e-4fced52b5a86", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1793837736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e238d5c1668540d2a9fffd3fd832d9b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1888ca08-3a", "ovs_interfaceid": "1888ca08-3a84-45c4-9fe2-79295174318d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.443795] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.220s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.444178] env[63021]: INFO nova.compute.manager [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Migrating [ 1035.518511] env[63021]: DEBUG oslo_vmware.api [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294225, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.538597] env[63021]: DEBUG oslo_concurrency.lockutils [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Releasing lock "refresh_cache-acc38ba0-515d-41d4-81dd-2b4374e033e7" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.538985] env[63021]: DEBUG nova.compute.manager [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Instance network_info: |[{"id": "1888ca08-3a84-45c4-9fe2-79295174318d", "address": "fa:16:3e:d9:b4:fb", "network": {"id": "81917a55-4e41-4898-9f4e-4fced52b5a86", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1793837736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e238d5c1668540d2a9fffd3fd832d9b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1888ca08-3a", "ovs_interfaceid": "1888ca08-3a84-45c4-9fe2-79295174318d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1035.539715] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d9:b4:fb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '535b175f-71d3-4226-81ab-ca253f27fedd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1888ca08-3a84-45c4-9fe2-79295174318d', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1035.547471] env[63021]: DEBUG oslo.service.loopingcall [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1035.547471] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1035.548276] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ed662dbf-96a1-41ef-8b36-afb52e4b4f88 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.568201] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1035.568201] env[63021]: value = "task-1294226" [ 1035.568201] env[63021]: _type = "Task" [ 1035.568201] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.575707] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294226, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.795255] env[63021]: DEBUG nova.compute.manager [req-571873da-e561-4bb0-881f-0cad9e9b617f req-92ebec69-c264-4da2-bc84-b90b49b0ae44 service nova] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Received event network-changed-1888ca08-3a84-45c4-9fe2-79295174318d {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1035.795609] env[63021]: DEBUG nova.compute.manager [req-571873da-e561-4bb0-881f-0cad9e9b617f req-92ebec69-c264-4da2-bc84-b90b49b0ae44 service nova] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Refreshing instance network info cache due to event network-changed-1888ca08-3a84-45c4-9fe2-79295174318d. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1035.795868] env[63021]: DEBUG oslo_concurrency.lockutils [req-571873da-e561-4bb0-881f-0cad9e9b617f req-92ebec69-c264-4da2-bc84-b90b49b0ae44 service nova] Acquiring lock "refresh_cache-acc38ba0-515d-41d4-81dd-2b4374e033e7" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.796031] env[63021]: DEBUG oslo_concurrency.lockutils [req-571873da-e561-4bb0-881f-0cad9e9b617f req-92ebec69-c264-4da2-bc84-b90b49b0ae44 service nova] Acquired lock "refresh_cache-acc38ba0-515d-41d4-81dd-2b4374e033e7" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.796271] env[63021]: DEBUG nova.network.neutron [req-571873da-e561-4bb0-881f-0cad9e9b617f req-92ebec69-c264-4da2-bc84-b90b49b0ae44 service nova] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Refreshing network info cache for port 1888ca08-3a84-45c4-9fe2-79295174318d {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1035.958583] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "refresh_cache-c328d2f7-3398-4f25-b11c-f464be7af8a1" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.958782] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquired lock "refresh_cache-c328d2f7-3398-4f25-b11c-f464be7af8a1" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.959373] env[63021]: DEBUG nova.network.neutron [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1036.018818] env[63021]: DEBUG oslo_vmware.api [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294225, 'name': ReconfigVM_Task, 'duration_secs': 0.697769} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.019130] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Reconfigured VM instance instance-00000063 to attach disk [datastore2] 63d5b0e4-0d61-4fa7-b193-2eaa171fc461/63d5b0e4-0d61-4fa7-b193-2eaa171fc461.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1036.019754] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f812acba-cdad-4a7c-8257-70c1faff3167 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.025970] env[63021]: DEBUG oslo_vmware.api [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1036.025970] env[63021]: value = "task-1294227" [ 1036.025970] env[63021]: _type = "Task" [ 1036.025970] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.033652] env[63021]: DEBUG oslo_vmware.api [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294227, 'name': Rename_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.078366] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294226, 'name': CreateVM_Task, 'duration_secs': 0.372365} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.078604] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1036.079416] env[63021]: DEBUG oslo_concurrency.lockutils [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1036.079622] env[63021]: DEBUG oslo_concurrency.lockutils [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.079983] env[63021]: DEBUG oslo_concurrency.lockutils [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1036.080635] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d000198-5e07-4030-b328-7207053b4d35 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.085693] env[63021]: DEBUG oslo_vmware.api [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1036.085693] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52568996-468c-3164-9abd-00e7ec11760b" [ 1036.085693] env[63021]: _type = "Task" [ 1036.085693] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.094178] env[63021]: DEBUG oslo_vmware.api [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52568996-468c-3164-9abd-00e7ec11760b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.324361] env[63021]: INFO nova.compute.manager [None req-141d343f-c79e-4d7b-a369-6f4b6b4741f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Unrescuing [ 1036.324361] env[63021]: DEBUG oslo_concurrency.lockutils [None req-141d343f-c79e-4d7b-a369-6f4b6b4741f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "refresh_cache-67f736d9-82ec-448e-8b72-84fdd704aa8d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1036.324361] env[63021]: DEBUG oslo_concurrency.lockutils [None req-141d343f-c79e-4d7b-a369-6f4b6b4741f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquired lock "refresh_cache-67f736d9-82ec-448e-8b72-84fdd704aa8d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.324361] env[63021]: DEBUG nova.network.neutron [None req-141d343f-c79e-4d7b-a369-6f4b6b4741f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1036.536836] env[63021]: DEBUG oslo_vmware.api [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294227, 'name': Rename_Task, 'duration_secs': 0.168685} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.537512] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1036.537512] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-77a6e721-d69c-4e62-ac03-50ac8e44040d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.543638] env[63021]: DEBUG oslo_vmware.api [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1036.543638] env[63021]: value = "task-1294228" [ 1036.543638] env[63021]: _type = "Task" [ 1036.543638] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.553967] env[63021]: DEBUG oslo_vmware.api [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294228, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.599026] env[63021]: DEBUG oslo_vmware.api [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52568996-468c-3164-9abd-00e7ec11760b, 'name': SearchDatastore_Task, 'duration_secs': 0.009495} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.599405] env[63021]: DEBUG oslo_concurrency.lockutils [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.599653] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1036.601513] env[63021]: DEBUG oslo_concurrency.lockutils [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1036.601513] env[63021]: DEBUG oslo_concurrency.lockutils [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.601513] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1036.601513] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-299e4719-056e-40c4-9e69-15e09b6230b9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.608505] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1036.608701] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1036.609443] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14aa33ba-96a0-4bba-b82c-bc81a4494bcc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.617804] env[63021]: DEBUG oslo_vmware.api [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1036.617804] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]526d16b8-1472-48fd-9d91-a54eb943744d" [ 1036.617804] env[63021]: _type = "Task" [ 1036.617804] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.626436] env[63021]: DEBUG oslo_vmware.api [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]526d16b8-1472-48fd-9d91-a54eb943744d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.701688] env[63021]: DEBUG nova.network.neutron [req-571873da-e561-4bb0-881f-0cad9e9b617f req-92ebec69-c264-4da2-bc84-b90b49b0ae44 service nova] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Updated VIF entry in instance network info cache for port 1888ca08-3a84-45c4-9fe2-79295174318d. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1036.701984] env[63021]: DEBUG nova.network.neutron [req-571873da-e561-4bb0-881f-0cad9e9b617f req-92ebec69-c264-4da2-bc84-b90b49b0ae44 service nova] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Updating instance_info_cache with network_info: [{"id": "1888ca08-3a84-45c4-9fe2-79295174318d", "address": "fa:16:3e:d9:b4:fb", "network": {"id": "81917a55-4e41-4898-9f4e-4fced52b5a86", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1793837736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e238d5c1668540d2a9fffd3fd832d9b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1888ca08-3a", "ovs_interfaceid": "1888ca08-3a84-45c4-9fe2-79295174318d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.711392] env[63021]: DEBUG nova.network.neutron [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Updating instance_info_cache with network_info: [{"id": "eff6f35a-0a13-4e0c-88b0-96b66b240f9b", "address": "fa:16:3e:27:3b:5e", "network": {"id": "9502e364-60e1-47ef-a1f7-f3e3ff45cf69", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-164717784-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5af87713ae6b477ca8a584aed287889d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeff6f35a-0a", "ovs_interfaceid": "eff6f35a-0a13-4e0c-88b0-96b66b240f9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.055599] env[63021]: DEBUG oslo_vmware.api [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294228, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.055742] env[63021]: DEBUG nova.network.neutron [None req-141d343f-c79e-4d7b-a369-6f4b6b4741f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Updating instance_info_cache with network_info: [{"id": "db333f88-b16f-439c-99fe-ca5884a09dc1", "address": "fa:16:3e:40:f8:6d", "network": {"id": "3aaeed35-f452-4144-a259-8f440f8a8bf4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-229941363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.192", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b672b1d7d7ad454ebfa3e2fd2421c754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb333f88-b1", "ovs_interfaceid": "db333f88-b16f-439c-99fe-ca5884a09dc1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.128127] env[63021]: DEBUG oslo_vmware.api [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]526d16b8-1472-48fd-9d91-a54eb943744d, 'name': SearchDatastore_Task, 'duration_secs': 0.01325} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.128910] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-269386ad-71df-4094-9179-5193bd8a0666 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.133807] env[63021]: DEBUG oslo_vmware.api [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1037.133807] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52a9e005-3d47-4f86-b996-0c5dfab7a7ae" [ 1037.133807] env[63021]: _type = "Task" [ 1037.133807] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.141842] env[63021]: DEBUG oslo_vmware.api [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52a9e005-3d47-4f86-b996-0c5dfab7a7ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.205239] env[63021]: DEBUG oslo_concurrency.lockutils [req-571873da-e561-4bb0-881f-0cad9e9b617f req-92ebec69-c264-4da2-bc84-b90b49b0ae44 service nova] Releasing lock "refresh_cache-acc38ba0-515d-41d4-81dd-2b4374e033e7" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.215115] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Releasing lock "refresh_cache-c328d2f7-3398-4f25-b11c-f464be7af8a1" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.555657] env[63021]: DEBUG oslo_vmware.api [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294228, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.559370] env[63021]: DEBUG oslo_concurrency.lockutils [None req-141d343f-c79e-4d7b-a369-6f4b6b4741f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Releasing lock "refresh_cache-67f736d9-82ec-448e-8b72-84fdd704aa8d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.560040] env[63021]: DEBUG nova.objects.instance [None req-141d343f-c79e-4d7b-a369-6f4b6b4741f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lazy-loading 'flavor' on Instance uuid 67f736d9-82ec-448e-8b72-84fdd704aa8d {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1037.645317] env[63021]: DEBUG oslo_vmware.api [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52a9e005-3d47-4f86-b996-0c5dfab7a7ae, 'name': SearchDatastore_Task, 'duration_secs': 0.012916} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.645618] env[63021]: DEBUG oslo_concurrency.lockutils [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.645914] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] acc38ba0-515d-41d4-81dd-2b4374e033e7/acc38ba0-515d-41d4-81dd-2b4374e033e7.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1037.646201] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ae5b39d2-4fef-451c-8456-c78ce729a885 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.653031] env[63021]: DEBUG oslo_vmware.api [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1037.653031] env[63021]: value = "task-1294229" [ 1037.653031] env[63021]: _type = "Task" [ 1037.653031] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.660663] env[63021]: DEBUG oslo_vmware.api [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294229, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.057979] env[63021]: DEBUG oslo_vmware.api [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294228, 'name': PowerOnVM_Task, 'duration_secs': 1.194749} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.059218] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1038.059218] env[63021]: INFO nova.compute.manager [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Took 10.32 seconds to spawn the instance on the hypervisor. [ 1038.059294] env[63021]: DEBUG nova.compute.manager [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1038.060220] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43f34855-eb86-4d36-8628-d2db98bd473f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.068229] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1992ce6-1ddc-423c-8b16-fb37743aa00d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.103959] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-141d343f-c79e-4d7b-a369-6f4b6b4741f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1038.104572] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9505fa3d-741a-4f9b-8fa1-c4a50c1aeb6a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.112029] env[63021]: DEBUG oslo_vmware.api [None req-141d343f-c79e-4d7b-a369-6f4b6b4741f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 1038.112029] env[63021]: value = "task-1294230" [ 1038.112029] env[63021]: _type = "Task" [ 1038.112029] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.121940] env[63021]: DEBUG oslo_vmware.api [None req-141d343f-c79e-4d7b-a369-6f4b6b4741f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294230, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.164727] env[63021]: DEBUG oslo_vmware.api [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294229, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.589947] env[63021]: INFO nova.compute.manager [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Took 16.25 seconds to build instance. [ 1038.626691] env[63021]: DEBUG oslo_vmware.api [None req-141d343f-c79e-4d7b-a369-6f4b6b4741f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294230, 'name': PowerOffVM_Task, 'duration_secs': 0.281829} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.626962] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-141d343f-c79e-4d7b-a369-6f4b6b4741f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1038.633387] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-141d343f-c79e-4d7b-a369-6f4b6b4741f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Reconfiguring VM instance instance-0000005b to detach disk 2002 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1038.636020] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-20c6d35a-d596-4658-97ec-b2f5ed24db59 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.652485] env[63021]: DEBUG oslo_vmware.api [None req-141d343f-c79e-4d7b-a369-6f4b6b4741f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 1038.652485] env[63021]: value = "task-1294231" [ 1038.652485] env[63021]: _type = "Task" [ 1038.652485] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.663287] env[63021]: DEBUG oslo_vmware.api [None req-141d343f-c79e-4d7b-a369-6f4b6b4741f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294231, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.666302] env[63021]: DEBUG oslo_vmware.api [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294229, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.668072} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.666618] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] acc38ba0-515d-41d4-81dd-2b4374e033e7/acc38ba0-515d-41d4-81dd-2b4374e033e7.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1038.666752] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1038.667079] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4244e6fd-a340-4fcd-a02c-e276c0c8a3d9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.672666] env[63021]: DEBUG oslo_vmware.api [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1038.672666] env[63021]: value = "task-1294232" [ 1038.672666] env[63021]: _type = "Task" [ 1038.672666] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.681661] env[63021]: DEBUG oslo_vmware.api [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294232, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.731656] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef364990-11cf-4d95-9201-6a198c7fd87e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.752021] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Updating instance 'c328d2f7-3398-4f25-b11c-f464be7af8a1' progress to 0 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1039.092942] env[63021]: DEBUG oslo_concurrency.lockutils [None req-fcf49901-6666-4369-8cb0-badaa2390b49 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "63d5b0e4-0d61-4fa7-b193-2eaa171fc461" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.765s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.163676] env[63021]: DEBUG oslo_vmware.api [None req-141d343f-c79e-4d7b-a369-6f4b6b4741f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294231, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.183208] env[63021]: DEBUG oslo_vmware.api [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294232, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.140832} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.183806] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1039.184650] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83676b23-eec5-4661-970e-370de41c0707 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.207703] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] acc38ba0-515d-41d4-81dd-2b4374e033e7/acc38ba0-515d-41d4-81dd-2b4374e033e7.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1039.207703] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff743992-8420-4804-902a-c468333e206f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.227843] env[63021]: DEBUG oslo_vmware.api [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1039.227843] env[63021]: value = "task-1294233" [ 1039.227843] env[63021]: _type = "Task" [ 1039.227843] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.236792] env[63021]: DEBUG oslo_vmware.api [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294233, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.258831] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1039.259123] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9121cf79-5e40-4d15-89aa-6c70ac81fe95 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.265042] env[63021]: DEBUG oslo_vmware.api [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1039.265042] env[63021]: value = "task-1294234" [ 1039.265042] env[63021]: _type = "Task" [ 1039.265042] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.273181] env[63021]: DEBUG oslo_vmware.api [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294234, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.605503] env[63021]: DEBUG oslo_concurrency.lockutils [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "63d5b0e4-0d61-4fa7-b193-2eaa171fc461" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.605808] env[63021]: DEBUG oslo_concurrency.lockutils [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "63d5b0e4-0d61-4fa7-b193-2eaa171fc461" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.606044] env[63021]: DEBUG oslo_concurrency.lockutils [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "63d5b0e4-0d61-4fa7-b193-2eaa171fc461-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.606262] env[63021]: DEBUG oslo_concurrency.lockutils [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "63d5b0e4-0d61-4fa7-b193-2eaa171fc461-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.606452] env[63021]: DEBUG oslo_concurrency.lockutils [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "63d5b0e4-0d61-4fa7-b193-2eaa171fc461-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.609664] env[63021]: INFO nova.compute.manager [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Terminating instance [ 1039.611824] env[63021]: DEBUG nova.compute.manager [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1039.612036] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1039.612913] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b93506b9-d2c4-42a6-ad20-708de17a4e24 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.621163] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1039.621297] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d8cc57e5-e283-40c3-ab43-4b779786d576 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.628131] env[63021]: DEBUG oslo_vmware.api [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1039.628131] env[63021]: value = "task-1294235" [ 1039.628131] env[63021]: _type = "Task" [ 1039.628131] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.636490] env[63021]: DEBUG oslo_vmware.api [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294235, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.662576] env[63021]: DEBUG oslo_vmware.api [None req-141d343f-c79e-4d7b-a369-6f4b6b4741f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294231, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.738366] env[63021]: DEBUG oslo_vmware.api [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294233, 'name': ReconfigVM_Task, 'duration_secs': 0.276088} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.739828] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Reconfigured VM instance instance-00000064 to attach disk [datastore2] acc38ba0-515d-41d4-81dd-2b4374e033e7/acc38ba0-515d-41d4-81dd-2b4374e033e7.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1039.739828] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fa947601-6f22-49ec-befb-2ac5659e3224 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.746908] env[63021]: DEBUG oslo_vmware.api [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1039.746908] env[63021]: value = "task-1294236" [ 1039.746908] env[63021]: _type = "Task" [ 1039.746908] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.754919] env[63021]: DEBUG oslo_vmware.api [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294236, 'name': Rename_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.775728] env[63021]: DEBUG oslo_vmware.api [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294234, 'name': PowerOffVM_Task, 'duration_secs': 0.354139} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.776049] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1039.776244] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Updating instance 'c328d2f7-3398-4f25-b11c-f464be7af8a1' progress to 17 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1040.140892] env[63021]: DEBUG oslo_vmware.api [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294235, 'name': PowerOffVM_Task, 'duration_secs': 0.248658} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.140892] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1040.140892] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1040.141120] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8eed07e2-11a4-4990-8e60-02acf00e02ee {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.164602] env[63021]: DEBUG oslo_vmware.api [None req-141d343f-c79e-4d7b-a369-6f4b6b4741f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294231, 'name': ReconfigVM_Task, 'duration_secs': 1.25233} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.165071] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-141d343f-c79e-4d7b-a369-6f4b6b4741f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Reconfigured VM instance instance-0000005b to detach disk 2002 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1040.165276] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-141d343f-c79e-4d7b-a369-6f4b6b4741f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1040.165540] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3168ca01-fbf7-4a95-900c-f2e887daf992 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.173021] env[63021]: DEBUG oslo_vmware.api [None req-141d343f-c79e-4d7b-a369-6f4b6b4741f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 1040.173021] env[63021]: value = "task-1294238" [ 1040.173021] env[63021]: _type = "Task" [ 1040.173021] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.180924] env[63021]: DEBUG oslo_vmware.api [None req-141d343f-c79e-4d7b-a369-6f4b6b4741f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294238, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.204919] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1040.205338] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Deleting contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1040.205564] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Deleting the datastore file [datastore2] 63d5b0e4-0d61-4fa7-b193-2eaa171fc461 {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1040.205842] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f0896981-681f-4d93-b459-80a0ced2078e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.213527] env[63021]: DEBUG oslo_vmware.api [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1040.213527] env[63021]: value = "task-1294239" [ 1040.213527] env[63021]: _type = "Task" [ 1040.213527] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.225018] env[63021]: DEBUG oslo_vmware.api [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294239, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.257042] env[63021]: DEBUG oslo_vmware.api [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294236, 'name': Rename_Task, 'duration_secs': 0.143045} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.257042] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1040.257618] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c2d70c47-dbc3-4849-a06a-12a4a745734c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.263872] env[63021]: DEBUG oslo_vmware.api [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1040.263872] env[63021]: value = "task-1294240" [ 1040.263872] env[63021]: _type = "Task" [ 1040.263872] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.271713] env[63021]: DEBUG oslo_vmware.api [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294240, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.283166] env[63021]: DEBUG nova.virt.hardware [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1040.283459] env[63021]: DEBUG nova.virt.hardware [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1040.283623] env[63021]: DEBUG nova.virt.hardware [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1040.283811] env[63021]: DEBUG nova.virt.hardware [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1040.283959] env[63021]: DEBUG nova.virt.hardware [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1040.284403] env[63021]: DEBUG nova.virt.hardware [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1040.284825] env[63021]: DEBUG nova.virt.hardware [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1040.285120] env[63021]: DEBUG nova.virt.hardware [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1040.285365] env[63021]: DEBUG nova.virt.hardware [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1040.285594] env[63021]: DEBUG nova.virt.hardware [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1040.285852] env[63021]: DEBUG nova.virt.hardware [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1040.291964] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9e1ff28f-7b9e-443f-ba41-316e73db02c6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.311977] env[63021]: DEBUG oslo_vmware.api [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1040.311977] env[63021]: value = "task-1294241" [ 1040.311977] env[63021]: _type = "Task" [ 1040.311977] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.323015] env[63021]: DEBUG oslo_vmware.api [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294241, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.686260] env[63021]: DEBUG oslo_vmware.api [None req-141d343f-c79e-4d7b-a369-6f4b6b4741f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294238, 'name': PowerOnVM_Task, 'duration_secs': 0.396518} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.687043] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-141d343f-c79e-4d7b-a369-6f4b6b4741f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1040.687398] env[63021]: DEBUG nova.compute.manager [None req-141d343f-c79e-4d7b-a369-6f4b6b4741f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1040.688261] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6403ab92-a59d-420b-89a3-e587d92d5fb6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.723990] env[63021]: DEBUG oslo_vmware.api [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294239, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.183868} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.724287] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1040.724480] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Deleted contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1040.724726] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1040.724948] env[63021]: INFO nova.compute.manager [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1040.725277] env[63021]: DEBUG oslo.service.loopingcall [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1040.725556] env[63021]: DEBUG nova.compute.manager [-] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1040.725684] env[63021]: DEBUG nova.network.neutron [-] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1040.774620] env[63021]: DEBUG oslo_vmware.api [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294240, 'name': PowerOnVM_Task, 'duration_secs': 0.48042} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.774620] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1040.774620] env[63021]: INFO nova.compute.manager [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Took 7.32 seconds to spawn the instance on the hypervisor. [ 1040.774620] env[63021]: DEBUG nova.compute.manager [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1040.776275] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6072e173-190c-4967-85ca-9c10e66ff0f4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.825502] env[63021]: DEBUG oslo_vmware.api [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294241, 'name': ReconfigVM_Task, 'duration_secs': 0.14826} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.825830] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Updating instance 'c328d2f7-3398-4f25-b11c-f464be7af8a1' progress to 33 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1041.011659] env[63021]: DEBUG nova.compute.manager [req-9a7e8410-7105-47cf-915e-f8e2b98945a0 req-8a4bcf21-3264-40a0-98c5-41579bde1d7f service nova] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Received event network-vif-deleted-952b66c6-27ca-4614-8875-8a94e770495a {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1041.011870] env[63021]: INFO nova.compute.manager [req-9a7e8410-7105-47cf-915e-f8e2b98945a0 req-8a4bcf21-3264-40a0-98c5-41579bde1d7f service nova] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Neutron deleted interface 952b66c6-27ca-4614-8875-8a94e770495a; detaching it from the instance and deleting it from the info cache [ 1041.012057] env[63021]: DEBUG nova.network.neutron [req-9a7e8410-7105-47cf-915e-f8e2b98945a0 req-8a4bcf21-3264-40a0-98c5-41579bde1d7f service nova] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.293905] env[63021]: INFO nova.compute.manager [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Took 15.48 seconds to build instance. [ 1041.331923] env[63021]: DEBUG nova.virt.hardware [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1041.332205] env[63021]: DEBUG nova.virt.hardware [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1041.332368] env[63021]: DEBUG nova.virt.hardware [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1041.332568] env[63021]: DEBUG nova.virt.hardware [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1041.332720] env[63021]: DEBUG nova.virt.hardware [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1041.332870] env[63021]: DEBUG nova.virt.hardware [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1041.333088] env[63021]: DEBUG nova.virt.hardware [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1041.333299] env[63021]: DEBUG nova.virt.hardware [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1041.333582] env[63021]: DEBUG nova.virt.hardware [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1041.333762] env[63021]: DEBUG nova.virt.hardware [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1041.333943] env[63021]: DEBUG nova.virt.hardware [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1041.339939] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Reconfiguring VM instance instance-00000045 to detach disk 2000 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1041.340663] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c9d3e2b-582b-4503-8e2f-6e278e2a512e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.359977] env[63021]: DEBUG oslo_vmware.api [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1041.359977] env[63021]: value = "task-1294242" [ 1041.359977] env[63021]: _type = "Task" [ 1041.359977] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.368691] env[63021]: DEBUG oslo_vmware.api [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294242, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.483522] env[63021]: DEBUG nova.network.neutron [-] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.516288] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e2b11574-9e01-42c2-9abf-7788109756dd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.525510] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-338b8a58-3aec-4ca6-b8c8-121b158813b3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.556444] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Acquiring lock "d32d9644-de8b-464e-ae16-7fa775a622ee" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.556801] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Lock "d32d9644-de8b-464e-ae16-7fa775a622ee" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.557962] env[63021]: DEBUG nova.compute.manager [req-9a7e8410-7105-47cf-915e-f8e2b98945a0 req-8a4bcf21-3264-40a0-98c5-41579bde1d7f service nova] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Detach interface failed, port_id=952b66c6-27ca-4614-8875-8a94e770495a, reason: Instance 63d5b0e4-0d61-4fa7-b193-2eaa171fc461 could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1041.689033] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d99e8b30-1b97-4e06-bc85-f3f78f8a36e2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.696191] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-8413ba9a-5bbe-495f-abda-aa98b341c874 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Suspending the VM {{(pid=63021) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1041.696191] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-1912718a-1354-4fea-834d-52401cfe5644 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.701763] env[63021]: DEBUG oslo_vmware.api [None req-8413ba9a-5bbe-495f-abda-aa98b341c874 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1041.701763] env[63021]: value = "task-1294243" [ 1041.701763] env[63021]: _type = "Task" [ 1041.701763] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.709698] env[63021]: DEBUG oslo_vmware.api [None req-8413ba9a-5bbe-495f-abda-aa98b341c874 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294243, 'name': SuspendVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.796641] env[63021]: DEBUG oslo_concurrency.lockutils [None req-54c52b2a-6515-490d-94fb-6af88e95da52 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "acc38ba0-515d-41d4-81dd-2b4374e033e7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.990s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.873565] env[63021]: DEBUG oslo_vmware.api [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294242, 'name': ReconfigVM_Task, 'duration_secs': 0.165931} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.873936] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Reconfigured VM instance instance-00000045 to detach disk 2000 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1041.874870] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daedf329-3642-40ee-a215-1ea6c4ba71f5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.899264] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] c328d2f7-3398-4f25-b11c-f464be7af8a1/c328d2f7-3398-4f25-b11c-f464be7af8a1.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1041.899723] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b6b49dcb-9c93-455f-849f-3f1a6e4f8036 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.917940] env[63021]: DEBUG oslo_vmware.api [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1041.917940] env[63021]: value = "task-1294244" [ 1041.917940] env[63021]: _type = "Task" [ 1041.917940] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.926102] env[63021]: DEBUG oslo_vmware.api [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294244, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.987417] env[63021]: INFO nova.compute.manager [-] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Took 1.26 seconds to deallocate network for instance. [ 1042.059821] env[63021]: DEBUG nova.compute.manager [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1042.211875] env[63021]: DEBUG oslo_vmware.api [None req-8413ba9a-5bbe-495f-abda-aa98b341c874 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294243, 'name': SuspendVM_Task} progress is 70%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.428822] env[63021]: DEBUG oslo_vmware.api [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294244, 'name': ReconfigVM_Task, 'duration_secs': 0.413954} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.429084] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Reconfigured VM instance instance-00000045 to attach disk [datastore1] c328d2f7-3398-4f25-b11c-f464be7af8a1/c328d2f7-3398-4f25-b11c-f464be7af8a1.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1042.429373] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Updating instance 'c328d2f7-3398-4f25-b11c-f464be7af8a1' progress to 50 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1042.493015] env[63021]: DEBUG oslo_concurrency.lockutils [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.493311] env[63021]: DEBUG oslo_concurrency.lockutils [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.493537] env[63021]: DEBUG nova.objects.instance [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lazy-loading 'resources' on Instance uuid 63d5b0e4-0d61-4fa7-b193-2eaa171fc461 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1042.613027] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.712453] env[63021]: DEBUG oslo_vmware.api [None req-8413ba9a-5bbe-495f-abda-aa98b341c874 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294243, 'name': SuspendVM_Task, 'duration_secs': 0.610006} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.712453] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-8413ba9a-5bbe-495f-abda-aa98b341c874 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Suspended the VM {{(pid=63021) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1042.712873] env[63021]: DEBUG nova.compute.manager [None req-8413ba9a-5bbe-495f-abda-aa98b341c874 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1042.713362] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ceabf53-09c3-4428-acfe-1fe8a8f4f891 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.936769] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3875ee3-1b36-438d-9ceb-aff8892b049d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.955531] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-044706b3-1aad-4425-a31d-3c6acd81b535 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.974316] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Updating instance 'c328d2f7-3398-4f25-b11c-f464be7af8a1' progress to 67 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1043.043054] env[63021]: DEBUG nova.compute.manager [req-6ed811d9-2fb3-421d-80bc-f07c65f3b671 req-4f9767aa-4737-457e-9982-08df273335e1 service nova] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Received event network-changed-db333f88-b16f-439c-99fe-ca5884a09dc1 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1043.043293] env[63021]: DEBUG nova.compute.manager [req-6ed811d9-2fb3-421d-80bc-f07c65f3b671 req-4f9767aa-4737-457e-9982-08df273335e1 service nova] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Refreshing instance network info cache due to event network-changed-db333f88-b16f-439c-99fe-ca5884a09dc1. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1043.043526] env[63021]: DEBUG oslo_concurrency.lockutils [req-6ed811d9-2fb3-421d-80bc-f07c65f3b671 req-4f9767aa-4737-457e-9982-08df273335e1 service nova] Acquiring lock "refresh_cache-67f736d9-82ec-448e-8b72-84fdd704aa8d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.043673] env[63021]: DEBUG oslo_concurrency.lockutils [req-6ed811d9-2fb3-421d-80bc-f07c65f3b671 req-4f9767aa-4737-457e-9982-08df273335e1 service nova] Acquired lock "refresh_cache-67f736d9-82ec-448e-8b72-84fdd704aa8d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.043837] env[63021]: DEBUG nova.network.neutron [req-6ed811d9-2fb3-421d-80bc-f07c65f3b671 req-4f9767aa-4737-457e-9982-08df273335e1 service nova] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Refreshing network info cache for port db333f88-b16f-439c-99fe-ca5884a09dc1 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1043.136590] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6e8b785-00a9-4785-be05-79550e1e995d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.144069] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42d6308d-6b30-4429-bbd0-184dbb2a7ddd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.172937] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c17414f-39ca-42e4-9d7b-13b135361f5d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.180627] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7914d4ad-8913-491d-b47f-9561caf9ef9a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.193813] env[63021]: DEBUG nova.compute.provider_tree [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1043.514700] env[63021]: DEBUG nova.network.neutron [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Port eff6f35a-0a13-4e0c-88b0-96b66b240f9b binding to destination host cpu-1 is already ACTIVE {{(pid=63021) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1043.696706] env[63021]: DEBUG nova.scheduler.client.report [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1043.756164] env[63021]: DEBUG nova.network.neutron [req-6ed811d9-2fb3-421d-80bc-f07c65f3b671 req-4f9767aa-4737-457e-9982-08df273335e1 service nova] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Updated VIF entry in instance network info cache for port db333f88-b16f-439c-99fe-ca5884a09dc1. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1043.756544] env[63021]: DEBUG nova.network.neutron [req-6ed811d9-2fb3-421d-80bc-f07c65f3b671 req-4f9767aa-4737-457e-9982-08df273335e1 service nova] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Updating instance_info_cache with network_info: [{"id": "db333f88-b16f-439c-99fe-ca5884a09dc1", "address": "fa:16:3e:40:f8:6d", "network": {"id": "3aaeed35-f452-4144-a259-8f440f8a8bf4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-229941363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.192", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b672b1d7d7ad454ebfa3e2fd2421c754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb333f88-b1", "ovs_interfaceid": "db333f88-b16f-439c-99fe-ca5884a09dc1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.052130] env[63021]: DEBUG oslo_concurrency.lockutils [None req-93dd89cc-2278-4281-bad7-2454c16f8266 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "acc38ba0-515d-41d4-81dd-2b4374e033e7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.052414] env[63021]: DEBUG oslo_concurrency.lockutils [None req-93dd89cc-2278-4281-bad7-2454c16f8266 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "acc38ba0-515d-41d4-81dd-2b4374e033e7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.052628] env[63021]: DEBUG oslo_concurrency.lockutils [None req-93dd89cc-2278-4281-bad7-2454c16f8266 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "acc38ba0-515d-41d4-81dd-2b4374e033e7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.052814] env[63021]: DEBUG oslo_concurrency.lockutils [None req-93dd89cc-2278-4281-bad7-2454c16f8266 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "acc38ba0-515d-41d4-81dd-2b4374e033e7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.052985] env[63021]: DEBUG oslo_concurrency.lockutils [None req-93dd89cc-2278-4281-bad7-2454c16f8266 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "acc38ba0-515d-41d4-81dd-2b4374e033e7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.055035] env[63021]: INFO nova.compute.manager [None req-93dd89cc-2278-4281-bad7-2454c16f8266 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Terminating instance [ 1044.056774] env[63021]: DEBUG nova.compute.manager [None req-93dd89cc-2278-4281-bad7-2454c16f8266 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1044.056981] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-93dd89cc-2278-4281-bad7-2454c16f8266 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1044.057853] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08fe8a98-fda1-4d1d-956b-100b4c8afcc5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.065750] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-93dd89cc-2278-4281-bad7-2454c16f8266 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1044.065976] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5327ae59-bd1d-4eb8-b6d0-7f5d5571ad55 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.179927] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-93dd89cc-2278-4281-bad7-2454c16f8266 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1044.180144] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-93dd89cc-2278-4281-bad7-2454c16f8266 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Deleting contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1044.180343] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-93dd89cc-2278-4281-bad7-2454c16f8266 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Deleting the datastore file [datastore2] acc38ba0-515d-41d4-81dd-2b4374e033e7 {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1044.180619] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1564f831-2c56-402f-b453-151b2f281850 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.187385] env[63021]: DEBUG oslo_vmware.api [None req-93dd89cc-2278-4281-bad7-2454c16f8266 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1044.187385] env[63021]: value = "task-1294246" [ 1044.187385] env[63021]: _type = "Task" [ 1044.187385] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.194729] env[63021]: DEBUG oslo_vmware.api [None req-93dd89cc-2278-4281-bad7-2454c16f8266 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294246, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.201584] env[63021]: DEBUG oslo_concurrency.lockutils [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.708s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.204089] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.591s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.205717] env[63021]: INFO nova.compute.claims [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1044.221544] env[63021]: INFO nova.scheduler.client.report [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Deleted allocations for instance 63d5b0e4-0d61-4fa7-b193-2eaa171fc461 [ 1044.259176] env[63021]: DEBUG oslo_concurrency.lockutils [req-6ed811d9-2fb3-421d-80bc-f07c65f3b671 req-4f9767aa-4737-457e-9982-08df273335e1 service nova] Releasing lock "refresh_cache-67f736d9-82ec-448e-8b72-84fdd704aa8d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.259484] env[63021]: DEBUG nova.compute.manager [req-6ed811d9-2fb3-421d-80bc-f07c65f3b671 req-4f9767aa-4737-457e-9982-08df273335e1 service nova] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Received event network-changed-db333f88-b16f-439c-99fe-ca5884a09dc1 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1044.259665] env[63021]: DEBUG nova.compute.manager [req-6ed811d9-2fb3-421d-80bc-f07c65f3b671 req-4f9767aa-4737-457e-9982-08df273335e1 service nova] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Refreshing instance network info cache due to event network-changed-db333f88-b16f-439c-99fe-ca5884a09dc1. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1044.259891] env[63021]: DEBUG oslo_concurrency.lockutils [req-6ed811d9-2fb3-421d-80bc-f07c65f3b671 req-4f9767aa-4737-457e-9982-08df273335e1 service nova] Acquiring lock "refresh_cache-67f736d9-82ec-448e-8b72-84fdd704aa8d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.260055] env[63021]: DEBUG oslo_concurrency.lockutils [req-6ed811d9-2fb3-421d-80bc-f07c65f3b671 req-4f9767aa-4737-457e-9982-08df273335e1 service nova] Acquired lock "refresh_cache-67f736d9-82ec-448e-8b72-84fdd704aa8d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.260230] env[63021]: DEBUG nova.network.neutron [req-6ed811d9-2fb3-421d-80bc-f07c65f3b671 req-4f9767aa-4737-457e-9982-08df273335e1 service nova] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Refreshing network info cache for port db333f88-b16f-439c-99fe-ca5884a09dc1 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1044.536582] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "c328d2f7-3398-4f25-b11c-f464be7af8a1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.536827] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "c328d2f7-3398-4f25-b11c-f464be7af8a1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.536991] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "c328d2f7-3398-4f25-b11c-f464be7af8a1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.697953] env[63021]: DEBUG oslo_vmware.api [None req-93dd89cc-2278-4281-bad7-2454c16f8266 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294246, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165914} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.698397] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-93dd89cc-2278-4281-bad7-2454c16f8266 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1044.698723] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-93dd89cc-2278-4281-bad7-2454c16f8266 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Deleted contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1044.699048] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-93dd89cc-2278-4281-bad7-2454c16f8266 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1044.699296] env[63021]: INFO nova.compute.manager [None req-93dd89cc-2278-4281-bad7-2454c16f8266 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Took 0.64 seconds to destroy the instance on the hypervisor. [ 1044.699601] env[63021]: DEBUG oslo.service.loopingcall [None req-93dd89cc-2278-4281-bad7-2454c16f8266 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1044.699886] env[63021]: DEBUG nova.compute.manager [-] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1044.700051] env[63021]: DEBUG nova.network.neutron [-] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1044.729792] env[63021]: DEBUG oslo_concurrency.lockutils [None req-44d0065e-8832-417b-bee8-90c931e7cbb7 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "63d5b0e4-0d61-4fa7-b193-2eaa171fc461" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.124s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.071258] env[63021]: DEBUG nova.compute.manager [req-6b841260-0dff-493c-89bc-033afc15dafc req-48bda91b-3f25-4c7e-b0ac-05d42889889e service nova] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Received event network-vif-deleted-1888ca08-3a84-45c4-9fe2-79295174318d {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1045.071688] env[63021]: INFO nova.compute.manager [req-6b841260-0dff-493c-89bc-033afc15dafc req-48bda91b-3f25-4c7e-b0ac-05d42889889e service nova] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Neutron deleted interface 1888ca08-3a84-45c4-9fe2-79295174318d; detaching it from the instance and deleting it from the info cache [ 1045.071688] env[63021]: DEBUG nova.network.neutron [req-6b841260-0dff-493c-89bc-033afc15dafc req-48bda91b-3f25-4c7e-b0ac-05d42889889e service nova] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.251472] env[63021]: DEBUG nova.network.neutron [req-6ed811d9-2fb3-421d-80bc-f07c65f3b671 req-4f9767aa-4737-457e-9982-08df273335e1 service nova] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Updated VIF entry in instance network info cache for port db333f88-b16f-439c-99fe-ca5884a09dc1. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1045.251978] env[63021]: DEBUG nova.network.neutron [req-6ed811d9-2fb3-421d-80bc-f07c65f3b671 req-4f9767aa-4737-457e-9982-08df273335e1 service nova] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Updating instance_info_cache with network_info: [{"id": "db333f88-b16f-439c-99fe-ca5884a09dc1", "address": "fa:16:3e:40:f8:6d", "network": {"id": "3aaeed35-f452-4144-a259-8f440f8a8bf4", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-229941363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.192", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b672b1d7d7ad454ebfa3e2fd2421c754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb333f88-b1", "ovs_interfaceid": "db333f88-b16f-439c-99fe-ca5884a09dc1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.361732] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93307f6f-8a3c-4eb9-a6f2-bc6c6973a34e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.370064] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b847ea19-c993-41db-a59b-a152254322bc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.400784] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e05e5f14-581a-404b-81df-105e15519505 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.408677] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71dc6d39-3b28-4000-9633-f3bcfa724558 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.422808] env[63021]: DEBUG nova.compute.provider_tree [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1045.494732] env[63021]: DEBUG nova.network.neutron [-] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.573720] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "refresh_cache-c328d2f7-3398-4f25-b11c-f464be7af8a1" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.573926] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquired lock "refresh_cache-c328d2f7-3398-4f25-b11c-f464be7af8a1" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.574120] env[63021]: DEBUG nova.network.neutron [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1045.575600] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8a760c91-daf3-48e0-ae4a-bcc8d1939ee3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.585634] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbef2803-2ec2-4b2f-aaaa-ff75bc20c769 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.612850] env[63021]: DEBUG nova.compute.manager [req-6b841260-0dff-493c-89bc-033afc15dafc req-48bda91b-3f25-4c7e-b0ac-05d42889889e service nova] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Detach interface failed, port_id=1888ca08-3a84-45c4-9fe2-79295174318d, reason: Instance acc38ba0-515d-41d4-81dd-2b4374e033e7 could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1045.756196] env[63021]: DEBUG oslo_concurrency.lockutils [req-6ed811d9-2fb3-421d-80bc-f07c65f3b671 req-4f9767aa-4737-457e-9982-08df273335e1 service nova] Releasing lock "refresh_cache-67f736d9-82ec-448e-8b72-84fdd704aa8d" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.891082] env[63021]: DEBUG oslo_concurrency.lockutils [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "e2e98322-fdda-45e0-b3ba-58349ab940be" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.891480] env[63021]: DEBUG oslo_concurrency.lockutils [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "e2e98322-fdda-45e0-b3ba-58349ab940be" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.925984] env[63021]: DEBUG nova.scheduler.client.report [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1045.997484] env[63021]: INFO nova.compute.manager [-] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Took 1.30 seconds to deallocate network for instance. [ 1046.309054] env[63021]: DEBUG nova.network.neutron [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Updating instance_info_cache with network_info: [{"id": "eff6f35a-0a13-4e0c-88b0-96b66b240f9b", "address": "fa:16:3e:27:3b:5e", "network": {"id": "9502e364-60e1-47ef-a1f7-f3e3ff45cf69", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-164717784-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5af87713ae6b477ca8a584aed287889d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeff6f35a-0a", "ovs_interfaceid": "eff6f35a-0a13-4e0c-88b0-96b66b240f9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.394040] env[63021]: DEBUG nova.compute.manager [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1046.431158] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.227s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.431676] env[63021]: DEBUG nova.compute.manager [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1046.505928] env[63021]: DEBUG oslo_concurrency.lockutils [None req-93dd89cc-2278-4281-bad7-2454c16f8266 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.506235] env[63021]: DEBUG oslo_concurrency.lockutils [None req-93dd89cc-2278-4281-bad7-2454c16f8266 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.506459] env[63021]: DEBUG nova.objects.instance [None req-93dd89cc-2278-4281-bad7-2454c16f8266 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lazy-loading 'resources' on Instance uuid acc38ba0-515d-41d4-81dd-2b4374e033e7 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1046.811676] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Releasing lock "refresh_cache-c328d2f7-3398-4f25-b11c-f464be7af8a1" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.915697] env[63021]: DEBUG oslo_concurrency.lockutils [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.936095] env[63021]: DEBUG nova.compute.utils [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1046.937627] env[63021]: DEBUG nova.compute.manager [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1046.937796] env[63021]: DEBUG nova.network.neutron [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1046.976310] env[63021]: DEBUG nova.policy [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6d2af434d8574ca2b480d408432179ed', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9bdaedef3a714f59864ef62dc41e55e7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 1047.155921] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79c02616-8ff8-4084-bee2-db91b6e3e35c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.164225] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75d93c66-e39e-4bc0-b25d-fb6cdae56039 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.194466] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5836922-13b9-4202-901e-4dfca7cd1931 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.201991] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d8b8617-88e0-489a-b6e5-0e7e25b29e94 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.215723] env[63021]: DEBUG nova.compute.provider_tree [None req-93dd89cc-2278-4281-bad7-2454c16f8266 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1047.316449] env[63021]: DEBUG nova.network.neutron [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Successfully created port: b7c97772-4a14-4bf5-9b07-980a1b3e92b6 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1047.336246] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d93dc30-2745-4ac5-bb6e-36092a640b0f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.356605] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-678a3647-b924-4e6a-ab98-fa092afbff07 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.363667] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Updating instance 'c328d2f7-3398-4f25-b11c-f464be7af8a1' progress to 83 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1047.440895] env[63021]: DEBUG nova.compute.manager [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1047.718758] env[63021]: DEBUG nova.scheduler.client.report [None req-93dd89cc-2278-4281-bad7-2454c16f8266 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1047.869887] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1047.870220] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-10e5fc9d-82a8-4a15-a816-e4ec577a1815 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.877929] env[63021]: DEBUG oslo_vmware.api [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1047.877929] env[63021]: value = "task-1294247" [ 1047.877929] env[63021]: _type = "Task" [ 1047.877929] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.885425] env[63021]: DEBUG oslo_vmware.api [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294247, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.224538] env[63021]: DEBUG oslo_concurrency.lockutils [None req-93dd89cc-2278-4281-bad7-2454c16f8266 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.718s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.227856] env[63021]: DEBUG oslo_concurrency.lockutils [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.311s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.229594] env[63021]: INFO nova.compute.claims [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1048.250374] env[63021]: INFO nova.scheduler.client.report [None req-93dd89cc-2278-4281-bad7-2454c16f8266 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Deleted allocations for instance acc38ba0-515d-41d4-81dd-2b4374e033e7 [ 1048.387568] env[63021]: DEBUG oslo_vmware.api [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294247, 'name': PowerOnVM_Task, 'duration_secs': 0.39706} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.387911] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1048.388062] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0c6d897e-656d-477d-ba71-c466b244b387 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Updating instance 'c328d2f7-3398-4f25-b11c-f464be7af8a1' progress to 100 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1048.423605] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1048.423835] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1048.423979] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Starting heal instance info cache {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1048.451921] env[63021]: DEBUG nova.compute.manager [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1048.477555] env[63021]: DEBUG nova.virt.hardware [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1048.477798] env[63021]: DEBUG nova.virt.hardware [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1048.477957] env[63021]: DEBUG nova.virt.hardware [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1048.478188] env[63021]: DEBUG nova.virt.hardware [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1048.478345] env[63021]: DEBUG nova.virt.hardware [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1048.478496] env[63021]: DEBUG nova.virt.hardware [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1048.478706] env[63021]: DEBUG nova.virt.hardware [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1048.478867] env[63021]: DEBUG nova.virt.hardware [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1048.479054] env[63021]: DEBUG nova.virt.hardware [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1048.479226] env[63021]: DEBUG nova.virt.hardware [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1048.479424] env[63021]: DEBUG nova.virt.hardware [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1048.480274] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0593916f-09a1-48c1-963c-828dba82a469 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.488201] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d906804-88ae-4477-bc32-55b3a843033b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.758258] env[63021]: DEBUG oslo_concurrency.lockutils [None req-93dd89cc-2278-4281-bad7-2454c16f8266 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "acc38ba0-515d-41d4-81dd-2b4374e033e7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.706s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.802796] env[63021]: DEBUG nova.compute.manager [req-2282310b-6b5c-4a77-b889-38a4bb8783ad req-b22ddc47-9a6b-4882-8ab3-dbf594feb5e6 service nova] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Received event network-vif-plugged-b7c97772-4a14-4bf5-9b07-980a1b3e92b6 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1048.803031] env[63021]: DEBUG oslo_concurrency.lockutils [req-2282310b-6b5c-4a77-b889-38a4bb8783ad req-b22ddc47-9a6b-4882-8ab3-dbf594feb5e6 service nova] Acquiring lock "d32d9644-de8b-464e-ae16-7fa775a622ee-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.803295] env[63021]: DEBUG oslo_concurrency.lockutils [req-2282310b-6b5c-4a77-b889-38a4bb8783ad req-b22ddc47-9a6b-4882-8ab3-dbf594feb5e6 service nova] Lock "d32d9644-de8b-464e-ae16-7fa775a622ee-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.803556] env[63021]: DEBUG oslo_concurrency.lockutils [req-2282310b-6b5c-4a77-b889-38a4bb8783ad req-b22ddc47-9a6b-4882-8ab3-dbf594feb5e6 service nova] Lock "d32d9644-de8b-464e-ae16-7fa775a622ee-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.803774] env[63021]: DEBUG nova.compute.manager [req-2282310b-6b5c-4a77-b889-38a4bb8783ad req-b22ddc47-9a6b-4882-8ab3-dbf594feb5e6 service nova] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] No waiting events found dispatching network-vif-plugged-b7c97772-4a14-4bf5-9b07-980a1b3e92b6 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1048.803955] env[63021]: WARNING nova.compute.manager [req-2282310b-6b5c-4a77-b889-38a4bb8783ad req-b22ddc47-9a6b-4882-8ab3-dbf594feb5e6 service nova] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Received unexpected event network-vif-plugged-b7c97772-4a14-4bf5-9b07-980a1b3e92b6 for instance with vm_state building and task_state spawning. [ 1049.073579] env[63021]: DEBUG nova.network.neutron [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Successfully updated port: b7c97772-4a14-4bf5-9b07-980a1b3e92b6 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1049.365028] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b1344f3-a543-4d8f-8cff-79fbe99774bb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.372507] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f8e4d84-d413-41be-a83e-d2328af80c4e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.408069] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e4e2323-1c38-4259-8621-b965016e2ebe {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.414879] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84c519f3-889a-402b-b71e-69df08ca7044 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.428459] env[63021]: DEBUG nova.compute.provider_tree [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1049.430810] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquiring lock "refresh_cache-c328d2f7-3398-4f25-b11c-f464be7af8a1" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1049.430966] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquired lock "refresh_cache-c328d2f7-3398-4f25-b11c-f464be7af8a1" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.431122] env[63021]: DEBUG nova.network.neutron [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Forcefully refreshing network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1049.576428] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Acquiring lock "refresh_cache-d32d9644-de8b-464e-ae16-7fa775a622ee" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1049.576576] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Acquired lock "refresh_cache-d32d9644-de8b-464e-ae16-7fa775a622ee" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.576735] env[63021]: DEBUG nova.network.neutron [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1049.933338] env[63021]: DEBUG nova.scheduler.client.report [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1050.107460] env[63021]: DEBUG nova.network.neutron [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1050.202783] env[63021]: DEBUG oslo_concurrency.lockutils [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "e940f625-f35e-459d-b8be-fa5916ce6efb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.203032] env[63021]: DEBUG oslo_concurrency.lockutils [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "e940f625-f35e-459d-b8be-fa5916ce6efb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.250577] env[63021]: DEBUG nova.network.neutron [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Updating instance_info_cache with network_info: [{"id": "b7c97772-4a14-4bf5-9b07-980a1b3e92b6", "address": "fa:16:3e:5e:a7:ec", "network": {"id": "a7fdca27-bbab-474e-9c33-0fba50028f45", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-976340152-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9bdaedef3a714f59864ef62dc41e55e7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "456bd8a2-0fb6-4b17-9d25-08e7995c5184", "external-id": "nsx-vlan-transportzone-65", "segmentation_id": 65, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7c97772-4a", "ovs_interfaceid": "b7c97772-4a14-4bf5-9b07-980a1b3e92b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.440075] env[63021]: DEBUG oslo_concurrency.lockutils [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.213s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.440425] env[63021]: DEBUG nova.compute.manager [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1050.705568] env[63021]: DEBUG nova.compute.manager [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1050.712147] env[63021]: DEBUG nova.network.neutron [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Updating instance_info_cache with network_info: [{"id": "eff6f35a-0a13-4e0c-88b0-96b66b240f9b", "address": "fa:16:3e:27:3b:5e", "network": {"id": "9502e364-60e1-47ef-a1f7-f3e3ff45cf69", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-164717784-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5af87713ae6b477ca8a584aed287889d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeff6f35a-0a", "ovs_interfaceid": "eff6f35a-0a13-4e0c-88b0-96b66b240f9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.753391] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Releasing lock "refresh_cache-d32d9644-de8b-464e-ae16-7fa775a622ee" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.753723] env[63021]: DEBUG nova.compute.manager [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Instance network_info: |[{"id": "b7c97772-4a14-4bf5-9b07-980a1b3e92b6", "address": "fa:16:3e:5e:a7:ec", "network": {"id": "a7fdca27-bbab-474e-9c33-0fba50028f45", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-976340152-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9bdaedef3a714f59864ef62dc41e55e7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "456bd8a2-0fb6-4b17-9d25-08e7995c5184", "external-id": "nsx-vlan-transportzone-65", "segmentation_id": 65, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7c97772-4a", "ovs_interfaceid": "b7c97772-4a14-4bf5-9b07-980a1b3e92b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1050.754149] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5e:a7:ec', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '456bd8a2-0fb6-4b17-9d25-08e7995c5184', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b7c97772-4a14-4bf5-9b07-980a1b3e92b6', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1050.761666] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Creating folder: Project (9bdaedef3a714f59864ef62dc41e55e7). Parent ref: group-v277447. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1050.762225] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ea48cc6a-2b00-4dc0-9418-48a9d40276fa {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.772034] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Created folder: Project (9bdaedef3a714f59864ef62dc41e55e7) in parent group-v277447. [ 1050.772238] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Creating folder: Instances. Parent ref: group-v277590. {{(pid=63021) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1050.772487] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3f956dec-d675-47c3-8e74-04bf3a69f6d2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.781377] env[63021]: INFO nova.virt.vmwareapi.vm_util [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Created folder: Instances in parent group-v277590. [ 1050.781608] env[63021]: DEBUG oslo.service.loopingcall [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1050.781795] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1050.782009] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f11947eb-9a35-4ce1-ae9e-ef48939fb5f6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.799837] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1050.799837] env[63021]: value = "task-1294250" [ 1050.799837] env[63021]: _type = "Task" [ 1050.799837] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.807591] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294250, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.828150] env[63021]: DEBUG nova.compute.manager [req-3a6ab7a7-d256-4985-9e5b-a178cfd7adc8 req-36da5601-8a43-4d8d-b6d5-a910410c1357 service nova] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Received event network-changed-b7c97772-4a14-4bf5-9b07-980a1b3e92b6 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1050.828381] env[63021]: DEBUG nova.compute.manager [req-3a6ab7a7-d256-4985-9e5b-a178cfd7adc8 req-36da5601-8a43-4d8d-b6d5-a910410c1357 service nova] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Refreshing instance network info cache due to event network-changed-b7c97772-4a14-4bf5-9b07-980a1b3e92b6. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1050.828711] env[63021]: DEBUG oslo_concurrency.lockutils [req-3a6ab7a7-d256-4985-9e5b-a178cfd7adc8 req-36da5601-8a43-4d8d-b6d5-a910410c1357 service nova] Acquiring lock "refresh_cache-d32d9644-de8b-464e-ae16-7fa775a622ee" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.828878] env[63021]: DEBUG oslo_concurrency.lockutils [req-3a6ab7a7-d256-4985-9e5b-a178cfd7adc8 req-36da5601-8a43-4d8d-b6d5-a910410c1357 service nova] Acquired lock "refresh_cache-d32d9644-de8b-464e-ae16-7fa775a622ee" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.829059] env[63021]: DEBUG nova.network.neutron [req-3a6ab7a7-d256-4985-9e5b-a178cfd7adc8 req-36da5601-8a43-4d8d-b6d5-a910410c1357 service nova] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Refreshing network info cache for port b7c97772-4a14-4bf5-9b07-980a1b3e92b6 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1050.945483] env[63021]: DEBUG nova.compute.utils [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1050.947023] env[63021]: DEBUG nova.compute.manager [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1050.947212] env[63021]: DEBUG nova.network.neutron [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1050.986610] env[63021]: DEBUG nova.policy [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ba16721213dd49a3b4e4990d7d319d1f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3c8959d2f1fe4706b7407c11e128b773', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 1051.178923] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dacdeb14-f674-47ef-b44e-57591f451403 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "c328d2f7-3398-4f25-b11c-f464be7af8a1" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.179186] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dacdeb14-f674-47ef-b44e-57591f451403 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "c328d2f7-3398-4f25-b11c-f464be7af8a1" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.179387] env[63021]: DEBUG nova.compute.manager [None req-dacdeb14-f674-47ef-b44e-57591f451403 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Going to confirm migration 3 {{(pid=63021) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1051.217067] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Releasing lock "refresh_cache-c328d2f7-3398-4f25-b11c-f464be7af8a1" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.217334] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Updated the network info_cache for instance {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1051.218099] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1051.218303] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1051.218493] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1051.218679] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1051.218859] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1051.219020] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1051.219192] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63021) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1051.219396] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1051.241891] env[63021]: DEBUG oslo_concurrency.lockutils [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.242187] env[63021]: DEBUG oslo_concurrency.lockutils [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.243694] env[63021]: INFO nova.compute.claims [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1051.247108] env[63021]: DEBUG nova.network.neutron [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Successfully created port: b9cfa391-5cd3-4e10-9ee1-547f1ee41769 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1051.310475] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294250, 'name': CreateVM_Task, 'duration_secs': 0.387679} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.310668] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1051.311423] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.311595] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.311921] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1051.312388] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f3f1400-b49c-4bc0-9d4d-5a357d7c9e80 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.316851] env[63021]: DEBUG oslo_vmware.api [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Waiting for the task: (returnval){ [ 1051.316851] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]520835de-4653-06be-bfa4-6d4de1e29260" [ 1051.316851] env[63021]: _type = "Task" [ 1051.316851] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.324878] env[63021]: DEBUG oslo_vmware.api [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]520835de-4653-06be-bfa4-6d4de1e29260, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.450532] env[63021]: DEBUG nova.compute.manager [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1051.523196] env[63021]: DEBUG nova.network.neutron [req-3a6ab7a7-d256-4985-9e5b-a178cfd7adc8 req-36da5601-8a43-4d8d-b6d5-a910410c1357 service nova] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Updated VIF entry in instance network info cache for port b7c97772-4a14-4bf5-9b07-980a1b3e92b6. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1051.523687] env[63021]: DEBUG nova.network.neutron [req-3a6ab7a7-d256-4985-9e5b-a178cfd7adc8 req-36da5601-8a43-4d8d-b6d5-a910410c1357 service nova] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Updating instance_info_cache with network_info: [{"id": "b7c97772-4a14-4bf5-9b07-980a1b3e92b6", "address": "fa:16:3e:5e:a7:ec", "network": {"id": "a7fdca27-bbab-474e-9c33-0fba50028f45", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-976340152-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9bdaedef3a714f59864ef62dc41e55e7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "456bd8a2-0fb6-4b17-9d25-08e7995c5184", "external-id": "nsx-vlan-transportzone-65", "segmentation_id": 65, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7c97772-4a", "ovs_interfaceid": "b7c97772-4a14-4bf5-9b07-980a1b3e92b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.722271] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.745221] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dacdeb14-f674-47ef-b44e-57591f451403 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "refresh_cache-c328d2f7-3398-4f25-b11c-f464be7af8a1" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.745430] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dacdeb14-f674-47ef-b44e-57591f451403 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquired lock "refresh_cache-c328d2f7-3398-4f25-b11c-f464be7af8a1" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.745610] env[63021]: DEBUG nova.network.neutron [None req-dacdeb14-f674-47ef-b44e-57591f451403 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1051.745797] env[63021]: DEBUG nova.objects.instance [None req-dacdeb14-f674-47ef-b44e-57591f451403 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lazy-loading 'info_cache' on Instance uuid c328d2f7-3398-4f25-b11c-f464be7af8a1 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1051.827483] env[63021]: DEBUG oslo_vmware.api [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]520835de-4653-06be-bfa4-6d4de1e29260, 'name': SearchDatastore_Task, 'duration_secs': 0.054434} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.827813] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.828079] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1051.828321] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.828469] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.828650] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1051.828922] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-43b069c0-977e-4d0f-a639-487b931779b7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.838071] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1051.838071] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1051.838475] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ce27c48-fdea-4097-bdaa-171e2e0b61d7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.843664] env[63021]: DEBUG oslo_vmware.api [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Waiting for the task: (returnval){ [ 1051.843664] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52338989-d6e2-f5da-9c4f-644badb87c8a" [ 1051.843664] env[63021]: _type = "Task" [ 1051.843664] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.850912] env[63021]: DEBUG oslo_vmware.api [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52338989-d6e2-f5da-9c4f-644badb87c8a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.026442] env[63021]: DEBUG oslo_concurrency.lockutils [req-3a6ab7a7-d256-4985-9e5b-a178cfd7adc8 req-36da5601-8a43-4d8d-b6d5-a910410c1357 service nova] Releasing lock "refresh_cache-d32d9644-de8b-464e-ae16-7fa775a622ee" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1052.355627] env[63021]: DEBUG oslo_vmware.api [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52338989-d6e2-f5da-9c4f-644badb87c8a, 'name': SearchDatastore_Task, 'duration_secs': 0.012048} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.356451] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0649e88f-45c0-47b0-9b02-656760bbab4c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.363208] env[63021]: DEBUG oslo_vmware.api [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Waiting for the task: (returnval){ [ 1052.363208] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52d75cfb-baff-1bae-f95d-d2ded942ca20" [ 1052.363208] env[63021]: _type = "Task" [ 1052.363208] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.370431] env[63021]: DEBUG oslo_vmware.api [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52d75cfb-baff-1bae-f95d-d2ded942ca20, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.380728] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b6e3feb-54a0-4d0f-aa41-59806413e96b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.386781] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-723e7b7e-1f8c-4600-bed1-ec4c7e66dabf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.417090] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db8d3dea-11f0-41e1-bac2-d6640ae2cc33 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.423535] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c48eda4f-04ea-4943-9e0c-1460770bb672 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.435884] env[63021]: DEBUG nova.compute.provider_tree [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1052.459300] env[63021]: DEBUG nova.compute.manager [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1052.485658] env[63021]: DEBUG nova.virt.hardware [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1052.485901] env[63021]: DEBUG nova.virt.hardware [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1052.486073] env[63021]: DEBUG nova.virt.hardware [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1052.486265] env[63021]: DEBUG nova.virt.hardware [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1052.486413] env[63021]: DEBUG nova.virt.hardware [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1052.486561] env[63021]: DEBUG nova.virt.hardware [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1052.486773] env[63021]: DEBUG nova.virt.hardware [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1052.486932] env[63021]: DEBUG nova.virt.hardware [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1052.487119] env[63021]: DEBUG nova.virt.hardware [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1052.487287] env[63021]: DEBUG nova.virt.hardware [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1052.487457] env[63021]: DEBUG nova.virt.hardware [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1052.488282] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f38f6c08-e0d8-4ab8-9b75-383888ff19f0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.495854] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04cb2a6f-beb9-4867-9359-1f864d677953 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.700321] env[63021]: DEBUG nova.network.neutron [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Successfully updated port: b9cfa391-5cd3-4e10-9ee1-547f1ee41769 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1052.857412] env[63021]: DEBUG nova.compute.manager [req-fdae88de-0f24-4905-a5cc-f10fc6f7e27a req-14614ecd-2fcb-4d66-bfee-b7399663f385 service nova] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Received event network-vif-plugged-b9cfa391-5cd3-4e10-9ee1-547f1ee41769 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1052.857412] env[63021]: DEBUG oslo_concurrency.lockutils [req-fdae88de-0f24-4905-a5cc-f10fc6f7e27a req-14614ecd-2fcb-4d66-bfee-b7399663f385 service nova] Acquiring lock "e2e98322-fdda-45e0-b3ba-58349ab940be-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.857412] env[63021]: DEBUG oslo_concurrency.lockutils [req-fdae88de-0f24-4905-a5cc-f10fc6f7e27a req-14614ecd-2fcb-4d66-bfee-b7399663f385 service nova] Lock "e2e98322-fdda-45e0-b3ba-58349ab940be-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.857412] env[63021]: DEBUG oslo_concurrency.lockutils [req-fdae88de-0f24-4905-a5cc-f10fc6f7e27a req-14614ecd-2fcb-4d66-bfee-b7399663f385 service nova] Lock "e2e98322-fdda-45e0-b3ba-58349ab940be-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.857412] env[63021]: DEBUG nova.compute.manager [req-fdae88de-0f24-4905-a5cc-f10fc6f7e27a req-14614ecd-2fcb-4d66-bfee-b7399663f385 service nova] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] No waiting events found dispatching network-vif-plugged-b9cfa391-5cd3-4e10-9ee1-547f1ee41769 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1052.857412] env[63021]: WARNING nova.compute.manager [req-fdae88de-0f24-4905-a5cc-f10fc6f7e27a req-14614ecd-2fcb-4d66-bfee-b7399663f385 service nova] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Received unexpected event network-vif-plugged-b9cfa391-5cd3-4e10-9ee1-547f1ee41769 for instance with vm_state building and task_state spawning. [ 1052.857412] env[63021]: DEBUG nova.compute.manager [req-fdae88de-0f24-4905-a5cc-f10fc6f7e27a req-14614ecd-2fcb-4d66-bfee-b7399663f385 service nova] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Received event network-changed-b9cfa391-5cd3-4e10-9ee1-547f1ee41769 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1052.857412] env[63021]: DEBUG nova.compute.manager [req-fdae88de-0f24-4905-a5cc-f10fc6f7e27a req-14614ecd-2fcb-4d66-bfee-b7399663f385 service nova] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Refreshing instance network info cache due to event network-changed-b9cfa391-5cd3-4e10-9ee1-547f1ee41769. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1052.857844] env[63021]: DEBUG oslo_concurrency.lockutils [req-fdae88de-0f24-4905-a5cc-f10fc6f7e27a req-14614ecd-2fcb-4d66-bfee-b7399663f385 service nova] Acquiring lock "refresh_cache-e2e98322-fdda-45e0-b3ba-58349ab940be" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1052.858178] env[63021]: DEBUG oslo_concurrency.lockutils [req-fdae88de-0f24-4905-a5cc-f10fc6f7e27a req-14614ecd-2fcb-4d66-bfee-b7399663f385 service nova] Acquired lock "refresh_cache-e2e98322-fdda-45e0-b3ba-58349ab940be" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1052.858489] env[63021]: DEBUG nova.network.neutron [req-fdae88de-0f24-4905-a5cc-f10fc6f7e27a req-14614ecd-2fcb-4d66-bfee-b7399663f385 service nova] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Refreshing network info cache for port b9cfa391-5cd3-4e10-9ee1-547f1ee41769 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1052.875038] env[63021]: DEBUG oslo_vmware.api [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52d75cfb-baff-1bae-f95d-d2ded942ca20, 'name': SearchDatastore_Task, 'duration_secs': 0.041777} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.875038] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1052.875038] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] d32d9644-de8b-464e-ae16-7fa775a622ee/d32d9644-de8b-464e-ae16-7fa775a622ee.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1052.875250] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6fce01d8-ca44-4b91-a4af-c806ded53e6b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.888549] env[63021]: DEBUG oslo_vmware.api [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Waiting for the task: (returnval){ [ 1052.888549] env[63021]: value = "task-1294251" [ 1052.888549] env[63021]: _type = "Task" [ 1052.888549] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.898383] env[63021]: DEBUG oslo_vmware.api [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Task: {'id': task-1294251, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.938916] env[63021]: DEBUG nova.scheduler.client.report [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1053.071600] env[63021]: DEBUG nova.network.neutron [None req-dacdeb14-f674-47ef-b44e-57591f451403 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Updating instance_info_cache with network_info: [{"id": "eff6f35a-0a13-4e0c-88b0-96b66b240f9b", "address": "fa:16:3e:27:3b:5e", "network": {"id": "9502e364-60e1-47ef-a1f7-f3e3ff45cf69", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-164717784-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5af87713ae6b477ca8a584aed287889d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeff6f35a-0a", "ovs_interfaceid": "eff6f35a-0a13-4e0c-88b0-96b66b240f9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.202529] env[63021]: DEBUG oslo_concurrency.lockutils [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "refresh_cache-e2e98322-fdda-45e0-b3ba-58349ab940be" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1053.389597] env[63021]: DEBUG nova.network.neutron [req-fdae88de-0f24-4905-a5cc-f10fc6f7e27a req-14614ecd-2fcb-4d66-bfee-b7399663f385 service nova] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1053.403630] env[63021]: DEBUG oslo_vmware.api [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Task: {'id': task-1294251, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.506048} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.403913] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] d32d9644-de8b-464e-ae16-7fa775a622ee/d32d9644-de8b-464e-ae16-7fa775a622ee.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1053.404162] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1053.404431] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-65c61db2-8e74-406d-bc6d-e9d9a547df57 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.410604] env[63021]: DEBUG oslo_vmware.api [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Waiting for the task: (returnval){ [ 1053.410604] env[63021]: value = "task-1294252" [ 1053.410604] env[63021]: _type = "Task" [ 1053.410604] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.417956] env[63021]: DEBUG oslo_vmware.api [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Task: {'id': task-1294252, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.445048] env[63021]: DEBUG oslo_concurrency.lockutils [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.203s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.445519] env[63021]: DEBUG nova.compute.manager [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1053.448121] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.726s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.448314] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.448463] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63021) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1053.449499] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8348db7-02d0-4076-9fb0-6b272f3ca541 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.457020] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01c5d88d-b7cc-40a5-a57a-b6d32d437073 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.470906] env[63021]: DEBUG nova.network.neutron [req-fdae88de-0f24-4905-a5cc-f10fc6f7e27a req-14614ecd-2fcb-4d66-bfee-b7399663f385 service nova] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.472448] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df7fb5c0-e776-47d2-bd93-a41a6ac29385 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.479252] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac6f592a-c280-4136-9b4e-7ee0707c2907 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.509135] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179969MB free_disk=138GB free_vcpus=48 pci_devices=None {{(pid=63021) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1053.509291] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.509489] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.573846] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dacdeb14-f674-47ef-b44e-57591f451403 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Releasing lock "refresh_cache-c328d2f7-3398-4f25-b11c-f464be7af8a1" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.574129] env[63021]: DEBUG nova.objects.instance [None req-dacdeb14-f674-47ef-b44e-57591f451403 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lazy-loading 'migration_context' on Instance uuid c328d2f7-3398-4f25-b11c-f464be7af8a1 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1053.923727] env[63021]: DEBUG oslo_vmware.api [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Task: {'id': task-1294252, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071636} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.924012] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1053.924839] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bbf154e-e324-4855-a14e-8c20289cb157 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.946678] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] d32d9644-de8b-464e-ae16-7fa775a622ee/d32d9644-de8b-464e-ae16-7fa775a622ee.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1053.947048] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-497e1684-4055-4441-8bb0-bda32f28df17 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.962401] env[63021]: DEBUG nova.compute.utils [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1053.963786] env[63021]: DEBUG nova.compute.manager [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1053.963955] env[63021]: DEBUG nova.network.neutron [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1053.971638] env[63021]: DEBUG oslo_vmware.api [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Waiting for the task: (returnval){ [ 1053.971638] env[63021]: value = "task-1294253" [ 1053.971638] env[63021]: _type = "Task" [ 1053.971638] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.975683] env[63021]: DEBUG oslo_concurrency.lockutils [req-fdae88de-0f24-4905-a5cc-f10fc6f7e27a req-14614ecd-2fcb-4d66-bfee-b7399663f385 service nova] Releasing lock "refresh_cache-e2e98322-fdda-45e0-b3ba-58349ab940be" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.976383] env[63021]: DEBUG oslo_concurrency.lockutils [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquired lock "refresh_cache-e2e98322-fdda-45e0-b3ba-58349ab940be" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.976571] env[63021]: DEBUG nova.network.neutron [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1053.981007] env[63021]: DEBUG oslo_vmware.api [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Task: {'id': task-1294253, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.020096] env[63021]: DEBUG nova.policy [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b427f7d1317341f38754a2b324f4f5aa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e238d5c1668540d2a9fffd3fd832d9b8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 1054.077332] env[63021]: DEBUG nova.objects.base [None req-dacdeb14-f674-47ef-b44e-57591f451403 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=63021) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1054.078432] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6d29ca5-c266-41fa-a40a-67115e8ff7e9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.097622] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b0806ebe-4487-4be0-99e6-f32c5a94c187 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.103200] env[63021]: DEBUG oslo_vmware.api [None req-dacdeb14-f674-47ef-b44e-57591f451403 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1054.103200] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]522d7530-573d-1a84-eba0-4d51f05c700e" [ 1054.103200] env[63021]: _type = "Task" [ 1054.103200] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.110676] env[63021]: DEBUG oslo_vmware.api [None req-dacdeb14-f674-47ef-b44e-57591f451403 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]522d7530-573d-1a84-eba0-4d51f05c700e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.356491] env[63021]: DEBUG nova.network.neutron [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Successfully created port: d2876adc-ed3c-4540-b06d-fd179ecd2445 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1054.467710] env[63021]: DEBUG nova.compute.manager [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1054.484179] env[63021]: DEBUG oslo_vmware.api [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Task: {'id': task-1294253, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.519326] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Applying migration context for instance c328d2f7-3398-4f25-b11c-f464be7af8a1 as it has an incoming, in-progress migration 3d2ed986-6753-426d-b8c0-b7b9ad33739e. Migration status is confirming {{(pid=63021) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1054.520642] env[63021]: INFO nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Updating resource usage from migration 3d2ed986-6753-426d-b8c0-b7b9ad33739e [ 1054.523859] env[63021]: DEBUG nova.network.neutron [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1054.545401] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance ae103118-bddf-46fe-90b9-98b60952ebba actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1054.545557] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 418a394b-dbb2-47a2-96a9-a1caa4c4bb51 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1054.545678] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 67f736d9-82ec-448e-8b72-84fdd704aa8d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1054.545796] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 129d8c8a-b869-422d-8e7f-cc3a9400021a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1054.545909] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 8d9387b8-6a84-4ada-8ac7-4ffd922b675a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1054.546034] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance d03c47da-7e73-4474-ad9b-07d27629dcae actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1054.546176] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Migration 3d2ed986-6753-426d-b8c0-b7b9ad33739e is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1054.546301] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance c328d2f7-3398-4f25-b11c-f464be7af8a1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1054.546417] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance d32d9644-de8b-464e-ae16-7fa775a622ee actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1054.546528] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance e2e98322-fdda-45e0-b3ba-58349ab940be actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1054.546637] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance e940f625-f35e-459d-b8be-fa5916ce6efb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1054.546832] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=63021) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1054.546965] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2688MB phys_disk=200GB used_disk=11GB total_vcpus=48 used_vcpus=11 pci_stats=[] {{(pid=63021) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1054.613291] env[63021]: DEBUG oslo_vmware.api [None req-dacdeb14-f674-47ef-b44e-57591f451403 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]522d7530-573d-1a84-eba0-4d51f05c700e, 'name': SearchDatastore_Task, 'duration_secs': 0.01275} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.615838] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dacdeb14-f674-47ef-b44e-57591f451403 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.669220] env[63021]: DEBUG nova.network.neutron [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Updating instance_info_cache with network_info: [{"id": "b9cfa391-5cd3-4e10-9ee1-547f1ee41769", "address": "fa:16:3e:e4:6e:90", "network": {"id": "0c0dd1a7-6687-45c5-82b1-33f34bdbb848", "bridge": "br-int", "label": "tempest-ServersTestJSON-523240467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c8959d2f1fe4706b7407c11e128b773", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffc811e4-0e4e-4748-8c7e-b3f14ccbd42d", "external-id": "nsx-vlan-transportzone-404", "segmentation_id": 404, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9cfa391-5c", "ovs_interfaceid": "b9cfa391-5cd3-4e10-9ee1-547f1ee41769", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.696596] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e7d800e-c490-490b-a6a9-2c8d9e7ed201 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.706116] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aafb688d-77c8-4ff5-8da8-eff5dd51dd41 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.735444] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63bcf8a6-1c8b-4598-b6f3-4d2cb88416f2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.742463] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f24961ea-658b-42f5-bee2-fdf9ff40972d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.755058] env[63021]: DEBUG nova.compute.provider_tree [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1054.983423] env[63021]: DEBUG oslo_vmware.api [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Task: {'id': task-1294253, 'name': ReconfigVM_Task, 'duration_secs': 0.74375} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.983701] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Reconfigured VM instance instance-00000065 to attach disk [datastore1] d32d9644-de8b-464e-ae16-7fa775a622ee/d32d9644-de8b-464e-ae16-7fa775a622ee.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1054.984331] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-373421c1-f6e0-46ed-8777-8ea3be0df2cb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.990786] env[63021]: DEBUG oslo_vmware.api [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Waiting for the task: (returnval){ [ 1054.990786] env[63021]: value = "task-1294254" [ 1054.990786] env[63021]: _type = "Task" [ 1054.990786] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.002412] env[63021]: DEBUG oslo_vmware.api [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Task: {'id': task-1294254, 'name': Rename_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.173979] env[63021]: DEBUG oslo_concurrency.lockutils [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Releasing lock "refresh_cache-e2e98322-fdda-45e0-b3ba-58349ab940be" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.175058] env[63021]: DEBUG nova.compute.manager [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Instance network_info: |[{"id": "b9cfa391-5cd3-4e10-9ee1-547f1ee41769", "address": "fa:16:3e:e4:6e:90", "network": {"id": "0c0dd1a7-6687-45c5-82b1-33f34bdbb848", "bridge": "br-int", "label": "tempest-ServersTestJSON-523240467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c8959d2f1fe4706b7407c11e128b773", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffc811e4-0e4e-4748-8c7e-b3f14ccbd42d", "external-id": "nsx-vlan-transportzone-404", "segmentation_id": 404, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9cfa391-5c", "ovs_interfaceid": "b9cfa391-5cd3-4e10-9ee1-547f1ee41769", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1055.175223] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e4:6e:90', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ffc811e4-0e4e-4748-8c7e-b3f14ccbd42d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b9cfa391-5cd3-4e10-9ee1-547f1ee41769', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1055.182526] env[63021]: DEBUG oslo.service.loopingcall [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1055.182761] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1055.182994] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7bcbd0ba-6d6a-4d0a-97dc-7a2c7c14233e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.204618] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1055.204618] env[63021]: value = "task-1294255" [ 1055.204618] env[63021]: _type = "Task" [ 1055.204618] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.214069] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294255, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.259633] env[63021]: DEBUG nova.scheduler.client.report [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1055.483200] env[63021]: DEBUG nova.compute.manager [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1055.499866] env[63021]: DEBUG oslo_vmware.api [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Task: {'id': task-1294254, 'name': Rename_Task, 'duration_secs': 0.137326} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.500285] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1055.500405] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c98e1247-ac5c-4b4a-a799-6a38957f7b47 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.508074] env[63021]: DEBUG oslo_vmware.api [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Waiting for the task: (returnval){ [ 1055.508074] env[63021]: value = "task-1294256" [ 1055.508074] env[63021]: _type = "Task" [ 1055.508074] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.510178] env[63021]: DEBUG nova.virt.hardware [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1055.510376] env[63021]: DEBUG nova.virt.hardware [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1055.510534] env[63021]: DEBUG nova.virt.hardware [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1055.510723] env[63021]: DEBUG nova.virt.hardware [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1055.510872] env[63021]: DEBUG nova.virt.hardware [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1055.511027] env[63021]: DEBUG nova.virt.hardware [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1055.511267] env[63021]: DEBUG nova.virt.hardware [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1055.511460] env[63021]: DEBUG nova.virt.hardware [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1055.511636] env[63021]: DEBUG nova.virt.hardware [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1055.511801] env[63021]: DEBUG nova.virt.hardware [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1055.511973] env[63021]: DEBUG nova.virt.hardware [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1055.512890] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c016de20-3485-451a-b32a-5309db18b437 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.525431] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44b430b9-1178-4e2b-8a1c-415d63ef3501 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.529110] env[63021]: DEBUG oslo_vmware.api [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Task: {'id': task-1294256, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.715353] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294255, 'name': CreateVM_Task, 'duration_secs': 0.344141} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.715563] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1055.719024] env[63021]: DEBUG oslo_concurrency.lockutils [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1055.719024] env[63021]: DEBUG oslo_concurrency.lockutils [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.719024] env[63021]: DEBUG oslo_concurrency.lockutils [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1055.719024] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41d8ed8a-ec96-4640-8bc7-e1426b1a98e0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.721931] env[63021]: DEBUG oslo_vmware.api [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1055.721931] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5266dc51-7ab7-68a1-d326-774f83b47b66" [ 1055.721931] env[63021]: _type = "Task" [ 1055.721931] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.729676] env[63021]: DEBUG oslo_vmware.api [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5266dc51-7ab7-68a1-d326-774f83b47b66, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.735903] env[63021]: DEBUG nova.compute.manager [req-22661ef0-81b3-49c5-bfc7-d7336cf45957 req-011057b2-a37d-4a64-9137-c904fc0ab436 service nova] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Received event network-vif-plugged-d2876adc-ed3c-4540-b06d-fd179ecd2445 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1055.736122] env[63021]: DEBUG oslo_concurrency.lockutils [req-22661ef0-81b3-49c5-bfc7-d7336cf45957 req-011057b2-a37d-4a64-9137-c904fc0ab436 service nova] Acquiring lock "e940f625-f35e-459d-b8be-fa5916ce6efb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.736357] env[63021]: DEBUG oslo_concurrency.lockutils [req-22661ef0-81b3-49c5-bfc7-d7336cf45957 req-011057b2-a37d-4a64-9137-c904fc0ab436 service nova] Lock "e940f625-f35e-459d-b8be-fa5916ce6efb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.736495] env[63021]: DEBUG oslo_concurrency.lockutils [req-22661ef0-81b3-49c5-bfc7-d7336cf45957 req-011057b2-a37d-4a64-9137-c904fc0ab436 service nova] Lock "e940f625-f35e-459d-b8be-fa5916ce6efb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.736665] env[63021]: DEBUG nova.compute.manager [req-22661ef0-81b3-49c5-bfc7-d7336cf45957 req-011057b2-a37d-4a64-9137-c904fc0ab436 service nova] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] No waiting events found dispatching network-vif-plugged-d2876adc-ed3c-4540-b06d-fd179ecd2445 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1055.736829] env[63021]: WARNING nova.compute.manager [req-22661ef0-81b3-49c5-bfc7-d7336cf45957 req-011057b2-a37d-4a64-9137-c904fc0ab436 service nova] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Received unexpected event network-vif-plugged-d2876adc-ed3c-4540-b06d-fd179ecd2445 for instance with vm_state building and task_state spawning. [ 1055.766884] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63021) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1055.767081] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.258s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.767349] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dacdeb14-f674-47ef-b44e-57591f451403 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 1.152s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.822968] env[63021]: DEBUG nova.network.neutron [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Successfully updated port: d2876adc-ed3c-4540-b06d-fd179ecd2445 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1056.021459] env[63021]: DEBUG oslo_vmware.api [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Task: {'id': task-1294256, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.232288] env[63021]: DEBUG oslo_vmware.api [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5266dc51-7ab7-68a1-d326-774f83b47b66, 'name': SearchDatastore_Task, 'duration_secs': 0.00954} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.232625] env[63021]: DEBUG oslo_concurrency.lockutils [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.232880] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1056.233186] env[63021]: DEBUG oslo_concurrency.lockutils [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1056.233364] env[63021]: DEBUG oslo_concurrency.lockutils [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.233575] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1056.233868] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-041d2b95-8cf0-4c57-a7e7-481dce7feb27 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.242350] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1056.242557] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1056.243270] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8725e7f2-4a5d-4be4-a8d2-7c89a8a500df {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.248306] env[63021]: DEBUG oslo_vmware.api [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1056.248306] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]527d4300-2754-0116-6d66-15c8f58040a6" [ 1056.248306] env[63021]: _type = "Task" [ 1056.248306] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.255789] env[63021]: DEBUG oslo_vmware.api [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]527d4300-2754-0116-6d66-15c8f58040a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.325256] env[63021]: DEBUG oslo_concurrency.lockutils [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "refresh_cache-e940f625-f35e-459d-b8be-fa5916ce6efb" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1056.325421] env[63021]: DEBUG oslo_concurrency.lockutils [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquired lock "refresh_cache-e940f625-f35e-459d-b8be-fa5916ce6efb" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.325583] env[63021]: DEBUG nova.network.neutron [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1056.400820] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd878ea1-20a5-4efe-9cb7-7d84b6873fa5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.409232] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ff9ab6c-9d7d-450e-b2f2-069a842798a3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.443520] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-433b9f62-5f03-44da-a184-8e06072da98a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.451282] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25929b21-d4c1-4c72-b32b-a5e8c3190b6a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.464106] env[63021]: DEBUG nova.compute.provider_tree [None req-dacdeb14-f674-47ef-b44e-57591f451403 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1056.524062] env[63021]: DEBUG oslo_vmware.api [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Task: {'id': task-1294256, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.758665] env[63021]: DEBUG oslo_vmware.api [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]527d4300-2754-0116-6d66-15c8f58040a6, 'name': SearchDatastore_Task, 'duration_secs': 0.013366} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.759513] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61c7b499-156f-493c-9380-1a9b065352f2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.764689] env[63021]: DEBUG oslo_vmware.api [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1056.764689] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52ffd799-08b4-34d6-7c0a-b7e2211d3720" [ 1056.764689] env[63021]: _type = "Task" [ 1056.764689] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.771824] env[63021]: DEBUG oslo_vmware.api [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52ffd799-08b4-34d6-7c0a-b7e2211d3720, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.856229] env[63021]: DEBUG nova.network.neutron [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1056.966899] env[63021]: DEBUG nova.scheduler.client.report [None req-dacdeb14-f674-47ef-b44e-57591f451403 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1056.989291] env[63021]: DEBUG nova.network.neutron [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Updating instance_info_cache with network_info: [{"id": "d2876adc-ed3c-4540-b06d-fd179ecd2445", "address": "fa:16:3e:a2:13:81", "network": {"id": "81917a55-4e41-4898-9f4e-4fced52b5a86", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1793837736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e238d5c1668540d2a9fffd3fd832d9b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2876adc-ed", "ovs_interfaceid": "d2876adc-ed3c-4540-b06d-fd179ecd2445", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.023146] env[63021]: DEBUG oslo_vmware.api [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Task: {'id': task-1294256, 'name': PowerOnVM_Task, 'duration_secs': 1.064211} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.023441] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1057.023645] env[63021]: INFO nova.compute.manager [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Took 8.57 seconds to spawn the instance on the hypervisor. [ 1057.023826] env[63021]: DEBUG nova.compute.manager [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1057.024620] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a94d7e7-b121-45af-a88f-4a64691fe1c2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.274888] env[63021]: DEBUG oslo_vmware.api [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52ffd799-08b4-34d6-7c0a-b7e2211d3720, 'name': SearchDatastore_Task, 'duration_secs': 0.010025} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.275149] env[63021]: DEBUG oslo_concurrency.lockutils [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.275423] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] e2e98322-fdda-45e0-b3ba-58349ab940be/e2e98322-fdda-45e0-b3ba-58349ab940be.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1057.275698] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-12e7e93d-072b-4ef8-9999-4f6a5f956132 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.281352] env[63021]: DEBUG oslo_vmware.api [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1057.281352] env[63021]: value = "task-1294257" [ 1057.281352] env[63021]: _type = "Task" [ 1057.281352] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.289826] env[63021]: DEBUG oslo_vmware.api [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294257, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.492200] env[63021]: DEBUG oslo_concurrency.lockutils [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Releasing lock "refresh_cache-e940f625-f35e-459d-b8be-fa5916ce6efb" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.492525] env[63021]: DEBUG nova.compute.manager [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Instance network_info: |[{"id": "d2876adc-ed3c-4540-b06d-fd179ecd2445", "address": "fa:16:3e:a2:13:81", "network": {"id": "81917a55-4e41-4898-9f4e-4fced52b5a86", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1793837736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e238d5c1668540d2a9fffd3fd832d9b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2876adc-ed", "ovs_interfaceid": "d2876adc-ed3c-4540-b06d-fd179ecd2445", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1057.493203] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a2:13:81', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '535b175f-71d3-4226-81ab-ca253f27fedd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd2876adc-ed3c-4540-b06d-fd179ecd2445', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1057.500772] env[63021]: DEBUG oslo.service.loopingcall [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1057.501327] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1057.501562] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c6c75273-3937-41c0-8202-c8547abeded6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.522314] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1057.522314] env[63021]: value = "task-1294258" [ 1057.522314] env[63021]: _type = "Task" [ 1057.522314] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.533675] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294258, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.540727] env[63021]: INFO nova.compute.manager [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Took 14.97 seconds to build instance. [ 1057.789414] env[63021]: DEBUG nova.compute.manager [req-9d85b479-f55e-451e-a5af-3edc6e37d203 req-7fcb3d3b-8514-4eb0-ac03-a49667a0eb32 service nova] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Received event network-changed-d2876adc-ed3c-4540-b06d-fd179ecd2445 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1057.789511] env[63021]: DEBUG nova.compute.manager [req-9d85b479-f55e-451e-a5af-3edc6e37d203 req-7fcb3d3b-8514-4eb0-ac03-a49667a0eb32 service nova] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Refreshing instance network info cache due to event network-changed-d2876adc-ed3c-4540-b06d-fd179ecd2445. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1057.789727] env[63021]: DEBUG oslo_concurrency.lockutils [req-9d85b479-f55e-451e-a5af-3edc6e37d203 req-7fcb3d3b-8514-4eb0-ac03-a49667a0eb32 service nova] Acquiring lock "refresh_cache-e940f625-f35e-459d-b8be-fa5916ce6efb" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.790069] env[63021]: DEBUG oslo_concurrency.lockutils [req-9d85b479-f55e-451e-a5af-3edc6e37d203 req-7fcb3d3b-8514-4eb0-ac03-a49667a0eb32 service nova] Acquired lock "refresh_cache-e940f625-f35e-459d-b8be-fa5916ce6efb" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.790069] env[63021]: DEBUG nova.network.neutron [req-9d85b479-f55e-451e-a5af-3edc6e37d203 req-7fcb3d3b-8514-4eb0-ac03-a49667a0eb32 service nova] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Refreshing network info cache for port d2876adc-ed3c-4540-b06d-fd179ecd2445 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1057.797120] env[63021]: DEBUG oslo_vmware.api [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294257, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.977721] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dacdeb14-f674-47ef-b44e-57591f451403 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.210s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.032653] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294258, 'name': CreateVM_Task, 'duration_secs': 0.423387} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.032799] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1058.033544] env[63021]: DEBUG oslo_concurrency.lockutils [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.033739] env[63021]: DEBUG oslo_concurrency.lockutils [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.034130] env[63021]: DEBUG oslo_concurrency.lockutils [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1058.034407] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe99e3b7-909c-457b-a440-a98c03e8c6b7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.039428] env[63021]: DEBUG oslo_vmware.api [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1058.039428] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52c1c67c-857e-2460-44c0-ad3214501b89" [ 1058.039428] env[63021]: _type = "Task" [ 1058.039428] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.042778] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dbcd12ba-5807-4718-b291-e793d2527544 tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Lock "d32d9644-de8b-464e-ae16-7fa775a622ee" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.486s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.048028] env[63021]: DEBUG oslo_vmware.api [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52c1c67c-857e-2460-44c0-ad3214501b89, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.294485] env[63021]: DEBUG oslo_vmware.api [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294257, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.529952} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.294805] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] e2e98322-fdda-45e0-b3ba-58349ab940be/e2e98322-fdda-45e0-b3ba-58349ab940be.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1058.295081] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1058.295711] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-23aef80b-42a4-44d6-9ed7-cc06a2d0da64 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.302395] env[63021]: DEBUG oslo_vmware.api [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1058.302395] env[63021]: value = "task-1294259" [ 1058.302395] env[63021]: _type = "Task" [ 1058.302395] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.310785] env[63021]: DEBUG oslo_vmware.api [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294259, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.537112] env[63021]: INFO nova.scheduler.client.report [None req-dacdeb14-f674-47ef-b44e-57591f451403 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Deleted allocation for migration 3d2ed986-6753-426d-b8c0-b7b9ad33739e [ 1058.552722] env[63021]: DEBUG nova.network.neutron [req-9d85b479-f55e-451e-a5af-3edc6e37d203 req-7fcb3d3b-8514-4eb0-ac03-a49667a0eb32 service nova] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Updated VIF entry in instance network info cache for port d2876adc-ed3c-4540-b06d-fd179ecd2445. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1058.553071] env[63021]: DEBUG nova.network.neutron [req-9d85b479-f55e-451e-a5af-3edc6e37d203 req-7fcb3d3b-8514-4eb0-ac03-a49667a0eb32 service nova] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Updating instance_info_cache with network_info: [{"id": "d2876adc-ed3c-4540-b06d-fd179ecd2445", "address": "fa:16:3e:a2:13:81", "network": {"id": "81917a55-4e41-4898-9f4e-4fced52b5a86", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1793837736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e238d5c1668540d2a9fffd3fd832d9b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2876adc-ed", "ovs_interfaceid": "d2876adc-ed3c-4540-b06d-fd179ecd2445", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.563029] env[63021]: DEBUG oslo_vmware.api [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52c1c67c-857e-2460-44c0-ad3214501b89, 'name': SearchDatastore_Task, 'duration_secs': 0.011177} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.563646] env[63021]: DEBUG oslo_concurrency.lockutils [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.563646] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1058.563792] env[63021]: DEBUG oslo_concurrency.lockutils [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.563924] env[63021]: DEBUG oslo_concurrency.lockutils [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.564168] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1058.564540] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-51da2867-e27c-46eb-a9a8-f365e3c6ce51 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.577272] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1058.578937] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1058.578937] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e36cf6f-1650-4749-a79f-4157ecd62d09 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.583884] env[63021]: DEBUG oslo_vmware.api [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1058.583884] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52c8fc4d-e476-fdf4-8cc8-604e5c595f91" [ 1058.583884] env[63021]: _type = "Task" [ 1058.583884] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.592348] env[63021]: DEBUG oslo_vmware.api [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52c8fc4d-e476-fdf4-8cc8-604e5c595f91, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.812374] env[63021]: DEBUG oslo_vmware.api [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294259, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067438} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.812499] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1058.814753] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c38937fa-2b1c-4bed-a529-51c3fcc2b1c5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.835622] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] e2e98322-fdda-45e0-b3ba-58349ab940be/e2e98322-fdda-45e0-b3ba-58349ab940be.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1058.836133] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ee586044-423c-463c-b831-0b9127465a81 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.856644] env[63021]: DEBUG oslo_vmware.api [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1058.856644] env[63021]: value = "task-1294260" [ 1058.856644] env[63021]: _type = "Task" [ 1058.856644] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.866614] env[63021]: DEBUG oslo_vmware.api [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294260, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.053525] env[63021]: DEBUG oslo_concurrency.lockutils [None req-dacdeb14-f674-47ef-b44e-57591f451403 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "c328d2f7-3398-4f25-b11c-f464be7af8a1" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.874s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.057564] env[63021]: DEBUG oslo_concurrency.lockutils [req-9d85b479-f55e-451e-a5af-3edc6e37d203 req-7fcb3d3b-8514-4eb0-ac03-a49667a0eb32 service nova] Releasing lock "refresh_cache-e940f625-f35e-459d-b8be-fa5916ce6efb" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.094595] env[63021]: DEBUG oslo_vmware.api [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52c8fc4d-e476-fdf4-8cc8-604e5c595f91, 'name': SearchDatastore_Task, 'duration_secs': 0.091006} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.095421] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28cdb6ee-6c62-419d-bb08-049d233da2be {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.102018] env[63021]: DEBUG oslo_vmware.api [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1059.102018] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5246f357-ee9c-2ee7-9aad-4311883bda6a" [ 1059.102018] env[63021]: _type = "Task" [ 1059.102018] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.112465] env[63021]: DEBUG oslo_vmware.api [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5246f357-ee9c-2ee7-9aad-4311883bda6a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.366564] env[63021]: DEBUG oslo_vmware.api [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294260, 'name': ReconfigVM_Task, 'duration_secs': 0.299742} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.366857] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Reconfigured VM instance instance-00000066 to attach disk [datastore1] e2e98322-fdda-45e0-b3ba-58349ab940be/e2e98322-fdda-45e0-b3ba-58349ab940be.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1059.367571] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0138b836-08ce-4648-b691-526163537c78 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.374239] env[63021]: DEBUG oslo_vmware.api [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1059.374239] env[63021]: value = "task-1294261" [ 1059.374239] env[63021]: _type = "Task" [ 1059.374239] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.382780] env[63021]: DEBUG oslo_vmware.api [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294261, 'name': Rename_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.612356] env[63021]: DEBUG oslo_vmware.api [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5246f357-ee9c-2ee7-9aad-4311883bda6a, 'name': SearchDatastore_Task, 'duration_secs': 0.018393} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.612693] env[63021]: DEBUG oslo_concurrency.lockutils [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.612929] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] e940f625-f35e-459d-b8be-fa5916ce6efb/e940f625-f35e-459d-b8be-fa5916ce6efb.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1059.613243] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0720536a-c188-4777-8c40-6fbab809eccd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.619485] env[63021]: DEBUG oslo_vmware.api [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1059.619485] env[63021]: value = "task-1294262" [ 1059.619485] env[63021]: _type = "Task" [ 1059.619485] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.626830] env[63021]: DEBUG oslo_vmware.api [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294262, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.662455] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e0a5a4d1-5720-42df-8a12-1d9857690211 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "c328d2f7-3398-4f25-b11c-f464be7af8a1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.662793] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e0a5a4d1-5720-42df-8a12-1d9857690211 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "c328d2f7-3398-4f25-b11c-f464be7af8a1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.663057] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e0a5a4d1-5720-42df-8a12-1d9857690211 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "c328d2f7-3398-4f25-b11c-f464be7af8a1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.663307] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e0a5a4d1-5720-42df-8a12-1d9857690211 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "c328d2f7-3398-4f25-b11c-f464be7af8a1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.663502] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e0a5a4d1-5720-42df-8a12-1d9857690211 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "c328d2f7-3398-4f25-b11c-f464be7af8a1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.665635] env[63021]: INFO nova.compute.manager [None req-e0a5a4d1-5720-42df-8a12-1d9857690211 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Terminating instance [ 1059.667453] env[63021]: DEBUG nova.compute.manager [None req-e0a5a4d1-5720-42df-8a12-1d9857690211 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1059.667652] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a5a4d1-5720-42df-8a12-1d9857690211 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1059.668477] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07637753-c16a-422d-a8a8-6a0f371ae230 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.675777] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0a5a4d1-5720-42df-8a12-1d9857690211 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1059.676013] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-369ac704-0811-45e7-ae94-14dfe94a2f97 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.681509] env[63021]: DEBUG oslo_vmware.api [None req-e0a5a4d1-5720-42df-8a12-1d9857690211 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1059.681509] env[63021]: value = "task-1294263" [ 1059.681509] env[63021]: _type = "Task" [ 1059.681509] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.690776] env[63021]: DEBUG oslo_vmware.api [None req-e0a5a4d1-5720-42df-8a12-1d9857690211 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294263, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.820544] env[63021]: DEBUG nova.compute.manager [req-8cf09f67-0408-467d-86b8-67a000985b41 req-87a224a7-df65-4164-9712-662a781cb49f service nova] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Received event network-changed-b7c97772-4a14-4bf5-9b07-980a1b3e92b6 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1059.820907] env[63021]: DEBUG nova.compute.manager [req-8cf09f67-0408-467d-86b8-67a000985b41 req-87a224a7-df65-4164-9712-662a781cb49f service nova] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Refreshing instance network info cache due to event network-changed-b7c97772-4a14-4bf5-9b07-980a1b3e92b6. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1059.821277] env[63021]: DEBUG oslo_concurrency.lockutils [req-8cf09f67-0408-467d-86b8-67a000985b41 req-87a224a7-df65-4164-9712-662a781cb49f service nova] Acquiring lock "refresh_cache-d32d9644-de8b-464e-ae16-7fa775a622ee" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.821522] env[63021]: DEBUG oslo_concurrency.lockutils [req-8cf09f67-0408-467d-86b8-67a000985b41 req-87a224a7-df65-4164-9712-662a781cb49f service nova] Acquired lock "refresh_cache-d32d9644-de8b-464e-ae16-7fa775a622ee" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.821820] env[63021]: DEBUG nova.network.neutron [req-8cf09f67-0408-467d-86b8-67a000985b41 req-87a224a7-df65-4164-9712-662a781cb49f service nova] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Refreshing network info cache for port b7c97772-4a14-4bf5-9b07-980a1b3e92b6 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1059.885482] env[63021]: DEBUG oslo_vmware.api [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294261, 'name': Rename_Task, 'duration_secs': 0.14569} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.885794] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1059.886220] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-14bf7c29-b182-4d2e-b835-8e5958cb069f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.893035] env[63021]: DEBUG oslo_vmware.api [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1059.893035] env[63021]: value = "task-1294264" [ 1059.893035] env[63021]: _type = "Task" [ 1059.893035] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.901460] env[63021]: DEBUG oslo_vmware.api [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294264, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.130775] env[63021]: DEBUG oslo_vmware.api [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294262, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.191846] env[63021]: DEBUG oslo_vmware.api [None req-e0a5a4d1-5720-42df-8a12-1d9857690211 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294263, 'name': PowerOffVM_Task, 'duration_secs': 0.241905} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.192173] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0a5a4d1-5720-42df-8a12-1d9857690211 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1060.192368] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a5a4d1-5720-42df-8a12-1d9857690211 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1060.192649] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bd421590-5bc6-4e99-8975-4720d4783a23 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.262996] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a5a4d1-5720-42df-8a12-1d9857690211 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1060.263260] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a5a4d1-5720-42df-8a12-1d9857690211 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1060.263470] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0a5a4d1-5720-42df-8a12-1d9857690211 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Deleting the datastore file [datastore1] c328d2f7-3398-4f25-b11c-f464be7af8a1 {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1060.263757] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d2dc2b9a-232c-4177-9ab1-d2925f0ae90b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.271159] env[63021]: DEBUG oslo_vmware.api [None req-e0a5a4d1-5720-42df-8a12-1d9857690211 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1060.271159] env[63021]: value = "task-1294266" [ 1060.271159] env[63021]: _type = "Task" [ 1060.271159] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.279667] env[63021]: DEBUG oslo_vmware.api [None req-e0a5a4d1-5720-42df-8a12-1d9857690211 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294266, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.402820] env[63021]: DEBUG oslo_vmware.api [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294264, 'name': PowerOnVM_Task, 'duration_secs': 0.470387} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.403123] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1060.403351] env[63021]: INFO nova.compute.manager [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Took 7.94 seconds to spawn the instance on the hypervisor. [ 1060.403536] env[63021]: DEBUG nova.compute.manager [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1060.404338] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7df13c6e-3a17-4319-9029-64fd1efb75ba {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.630710] env[63021]: DEBUG oslo_vmware.api [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294262, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.581962} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.631071] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] e940f625-f35e-459d-b8be-fa5916ce6efb/e940f625-f35e-459d-b8be-fa5916ce6efb.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1060.631226] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1060.631509] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f40bea9e-6f48-4f4b-99b0-eed7067699a6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.638372] env[63021]: DEBUG oslo_vmware.api [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1060.638372] env[63021]: value = "task-1294267" [ 1060.638372] env[63021]: _type = "Task" [ 1060.638372] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.646846] env[63021]: DEBUG oslo_vmware.api [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294267, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.780810] env[63021]: DEBUG oslo_vmware.api [None req-e0a5a4d1-5720-42df-8a12-1d9857690211 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294266, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.867744] env[63021]: DEBUG nova.network.neutron [req-8cf09f67-0408-467d-86b8-67a000985b41 req-87a224a7-df65-4164-9712-662a781cb49f service nova] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Updated VIF entry in instance network info cache for port b7c97772-4a14-4bf5-9b07-980a1b3e92b6. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1060.868819] env[63021]: DEBUG nova.network.neutron [req-8cf09f67-0408-467d-86b8-67a000985b41 req-87a224a7-df65-4164-9712-662a781cb49f service nova] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Updating instance_info_cache with network_info: [{"id": "b7c97772-4a14-4bf5-9b07-980a1b3e92b6", "address": "fa:16:3e:5e:a7:ec", "network": {"id": "a7fdca27-bbab-474e-9c33-0fba50028f45", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-976340152-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9bdaedef3a714f59864ef62dc41e55e7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "456bd8a2-0fb6-4b17-9d25-08e7995c5184", "external-id": "nsx-vlan-transportzone-65", "segmentation_id": 65, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7c97772-4a", "ovs_interfaceid": "b7c97772-4a14-4bf5-9b07-980a1b3e92b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.920447] env[63021]: INFO nova.compute.manager [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Took 14.02 seconds to build instance. [ 1061.149282] env[63021]: DEBUG oslo_vmware.api [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294267, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.204086} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.149645] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1061.150441] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04bbeca1-846c-4917-a809-746ba86462e7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.172067] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] e940f625-f35e-459d-b8be-fa5916ce6efb/e940f625-f35e-459d-b8be-fa5916ce6efb.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1061.172388] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-67642219-359d-4958-86d4-fb508466d0c1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.191763] env[63021]: DEBUG oslo_vmware.api [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1061.191763] env[63021]: value = "task-1294268" [ 1061.191763] env[63021]: _type = "Task" [ 1061.191763] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.199599] env[63021]: DEBUG oslo_vmware.api [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294268, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.280984] env[63021]: DEBUG oslo_vmware.api [None req-e0a5a4d1-5720-42df-8a12-1d9857690211 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294266, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.814957} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.281309] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0a5a4d1-5720-42df-8a12-1d9857690211 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1061.281529] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a5a4d1-5720-42df-8a12-1d9857690211 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1061.281747] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a5a4d1-5720-42df-8a12-1d9857690211 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1061.281944] env[63021]: INFO nova.compute.manager [None req-e0a5a4d1-5720-42df-8a12-1d9857690211 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Took 1.61 seconds to destroy the instance on the hypervisor. [ 1061.282236] env[63021]: DEBUG oslo.service.loopingcall [None req-e0a5a4d1-5720-42df-8a12-1d9857690211 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1061.282463] env[63021]: DEBUG nova.compute.manager [-] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1061.282576] env[63021]: DEBUG nova.network.neutron [-] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1061.370927] env[63021]: DEBUG oslo_concurrency.lockutils [req-8cf09f67-0408-467d-86b8-67a000985b41 req-87a224a7-df65-4164-9712-662a781cb49f service nova] Releasing lock "refresh_cache-d32d9644-de8b-464e-ae16-7fa775a622ee" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.422914] env[63021]: DEBUG oslo_concurrency.lockutils [None req-83e64712-e925-42df-b9b4-db0d0218a8ae tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "e2e98322-fdda-45e0-b3ba-58349ab940be" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.531s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.702067] env[63021]: DEBUG oslo_vmware.api [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294268, 'name': ReconfigVM_Task, 'duration_secs': 0.283621} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.702517] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Reconfigured VM instance instance-00000067 to attach disk [datastore1] e940f625-f35e-459d-b8be-fa5916ce6efb/e940f625-f35e-459d-b8be-fa5916ce6efb.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1061.703013] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-42cb4057-dd86-4ca7-a9c8-623ee33aadeb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.709555] env[63021]: DEBUG oslo_vmware.api [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1061.709555] env[63021]: value = "task-1294269" [ 1061.709555] env[63021]: _type = "Task" [ 1061.709555] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.722731] env[63021]: DEBUG oslo_vmware.api [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294269, 'name': Rename_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.744202] env[63021]: DEBUG oslo_concurrency.lockutils [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "a2d6db04-9293-4755-ad8c-2e46467f92cb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.744464] env[63021]: DEBUG oslo_concurrency.lockutils [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "a2d6db04-9293-4755-ad8c-2e46467f92cb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.853327] env[63021]: DEBUG nova.compute.manager [req-08e9dcb4-46d7-45f4-a249-5a1829fd8bdd req-63d5f1fe-a5a7-445d-9b77-c53108e3b745 service nova] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Received event network-vif-deleted-eff6f35a-0a13-4e0c-88b0-96b66b240f9b {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1061.853327] env[63021]: INFO nova.compute.manager [req-08e9dcb4-46d7-45f4-a249-5a1829fd8bdd req-63d5f1fe-a5a7-445d-9b77-c53108e3b745 service nova] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Neutron deleted interface eff6f35a-0a13-4e0c-88b0-96b66b240f9b; detaching it from the instance and deleting it from the info cache [ 1061.853574] env[63021]: DEBUG nova.network.neutron [req-08e9dcb4-46d7-45f4-a249-5a1829fd8bdd req-63d5f1fe-a5a7-445d-9b77-c53108e3b745 service nova] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.220831] env[63021]: DEBUG oslo_vmware.api [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294269, 'name': Rename_Task, 'duration_secs': 0.167094} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.221222] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1062.221485] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b0165025-b899-40ad-a5fe-d6b4b917af48 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.227565] env[63021]: DEBUG oslo_vmware.api [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1062.227565] env[63021]: value = "task-1294270" [ 1062.227565] env[63021]: _type = "Task" [ 1062.227565] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.235106] env[63021]: DEBUG oslo_vmware.api [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294270, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.237468] env[63021]: DEBUG nova.network.neutron [-] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.246772] env[63021]: DEBUG nova.compute.manager [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1062.356325] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-32ab939e-ded8-4557-8822-4c4c1348173c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.365164] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04b5afa0-85b0-4046-939b-f0e34c8bcbb3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.393687] env[63021]: DEBUG nova.compute.manager [req-08e9dcb4-46d7-45f4-a249-5a1829fd8bdd req-63d5f1fe-a5a7-445d-9b77-c53108e3b745 service nova] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Detach interface failed, port_id=eff6f35a-0a13-4e0c-88b0-96b66b240f9b, reason: Instance c328d2f7-3398-4f25-b11c-f464be7af8a1 could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1062.738266] env[63021]: DEBUG oslo_vmware.api [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294270, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.739760] env[63021]: INFO nova.compute.manager [-] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Took 1.46 seconds to deallocate network for instance. [ 1062.765193] env[63021]: DEBUG oslo_concurrency.lockutils [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.765439] env[63021]: DEBUG oslo_concurrency.lockutils [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.766849] env[63021]: INFO nova.compute.claims [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1063.125026] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3db7651d-23fa-4bc2-9db6-725c3ef3c5b3 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquiring lock "8d9387b8-6a84-4ada-8ac7-4ffd922b675a" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.125193] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3db7651d-23fa-4bc2-9db6-725c3ef3c5b3 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "8d9387b8-6a84-4ada-8ac7-4ffd922b675a" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.241015] env[63021]: DEBUG oslo_vmware.api [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294270, 'name': PowerOnVM_Task, 'duration_secs': 0.724397} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.241480] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1063.241813] env[63021]: INFO nova.compute.manager [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Took 7.76 seconds to spawn the instance on the hypervisor. [ 1063.242133] env[63021]: DEBUG nova.compute.manager [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1063.243325] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97dd2dcc-6c3c-4921-b647-4bf3397819c4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.248243] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e0a5a4d1-5720-42df-8a12-1d9857690211 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.629956] env[63021]: DEBUG nova.compute.utils [None req-3db7651d-23fa-4bc2-9db6-725c3ef3c5b3 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1063.764247] env[63021]: INFO nova.compute.manager [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Took 12.55 seconds to build instance. [ 1063.899545] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4c54f23-9e40-4e96-a5de-3474ecf6a54b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.908964] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9b900ac-fd89-489d-ab26-b531b91fc97a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.942326] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3d0bcb5-4d10-44c8-bfc9-7360aa9a3a41 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.950431] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3d94b45-fdac-4105-aff4-ba7eb8311346 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.964669] env[63021]: DEBUG nova.compute.provider_tree [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1064.133102] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3db7651d-23fa-4bc2-9db6-725c3ef3c5b3 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "8d9387b8-6a84-4ada-8ac7-4ffd922b675a" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.266597] env[63021]: DEBUG oslo_concurrency.lockutils [None req-400caf84-b5e8-45e4-8595-009decbd3bf0 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "e940f625-f35e-459d-b8be-fa5916ce6efb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.063s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.467516] env[63021]: DEBUG nova.scheduler.client.report [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1064.972233] env[63021]: DEBUG oslo_concurrency.lockutils [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.207s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.972760] env[63021]: DEBUG nova.compute.manager [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1064.975310] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e0a5a4d1-5720-42df-8a12-1d9857690211 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.727s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.975523] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e0a5a4d1-5720-42df-8a12-1d9857690211 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.992477] env[63021]: INFO nova.scheduler.client.report [None req-e0a5a4d1-5720-42df-8a12-1d9857690211 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Deleted allocations for instance c328d2f7-3398-4f25-b11c-f464be7af8a1 [ 1065.192359] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3db7651d-23fa-4bc2-9db6-725c3ef3c5b3 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquiring lock "8d9387b8-6a84-4ada-8ac7-4ffd922b675a" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.192575] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3db7651d-23fa-4bc2-9db6-725c3ef3c5b3 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "8d9387b8-6a84-4ada-8ac7-4ffd922b675a" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.192814] env[63021]: INFO nova.compute.manager [None req-3db7651d-23fa-4bc2-9db6-725c3ef3c5b3 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Attaching volume 29440f30-19f1-42a5-9503-a29373b566ff to /dev/sdb [ 1065.227843] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7021ac7a-4b62-47af-8252-a3a18aca8df4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.234882] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dfb6713-a80a-429a-b4b5-15c833ad32c9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.246946] env[63021]: DEBUG nova.virt.block_device [None req-3db7651d-23fa-4bc2-9db6-725c3ef3c5b3 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Updating existing volume attachment record: b2fe4b90-93d8-4aa0-8d32-61a00c9b528b {{(pid=63021) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1065.479729] env[63021]: DEBUG nova.compute.utils [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1065.481495] env[63021]: DEBUG nova.compute.manager [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1065.481770] env[63021]: DEBUG nova.network.neutron [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1065.501014] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e0a5a4d1-5720-42df-8a12-1d9857690211 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "c328d2f7-3398-4f25-b11c-f464be7af8a1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.838s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.540052] env[63021]: DEBUG nova.policy [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ba16721213dd49a3b4e4990d7d319d1f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3c8959d2f1fe4706b7407c11e128b773', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 1065.555808] env[63021]: DEBUG nova.compute.manager [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Stashing vm_state: active {{(pid=63021) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1065.792515] env[63021]: DEBUG nova.network.neutron [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Successfully created port: ef03a045-2fd4-4794-964b-ec6a72e8f7d2 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1065.985281] env[63021]: DEBUG nova.compute.manager [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1066.077606] env[63021]: DEBUG oslo_concurrency.lockutils [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.077894] env[63021]: DEBUG oslo_concurrency.lockutils [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.583761] env[63021]: INFO nova.compute.claims [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1066.858543] env[63021]: DEBUG oslo_concurrency.lockutils [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "4041dfbf-fa94-4ff8-9176-e7088cdf241f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.858778] env[63021]: DEBUG oslo_concurrency.lockutils [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "4041dfbf-fa94-4ff8-9176-e7088cdf241f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.999107] env[63021]: DEBUG nova.compute.manager [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1067.023685] env[63021]: DEBUG nova.virt.hardware [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1067.023931] env[63021]: DEBUG nova.virt.hardware [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1067.024108] env[63021]: DEBUG nova.virt.hardware [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1067.024312] env[63021]: DEBUG nova.virt.hardware [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1067.024465] env[63021]: DEBUG nova.virt.hardware [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1067.024613] env[63021]: DEBUG nova.virt.hardware [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1067.024822] env[63021]: DEBUG nova.virt.hardware [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1067.025046] env[63021]: DEBUG nova.virt.hardware [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1067.025241] env[63021]: DEBUG nova.virt.hardware [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1067.025414] env[63021]: DEBUG nova.virt.hardware [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1067.025591] env[63021]: DEBUG nova.virt.hardware [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1067.026461] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bfc9d4a-880d-4177-82e0-4afb6a68344d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.034544] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6115a5f2-15fc-404e-8b8f-870fd792fda6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.091161] env[63021]: INFO nova.compute.resource_tracker [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Updating resource usage from migration 863e91bb-e79b-4422-99bf-1ac76f99aafa [ 1067.237247] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c8a4773-31cc-48fd-900e-1edc9e0c99b8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.245033] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70407b08-189b-4938-84a5-73a264304661 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.274464] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc3dd255-471e-4ad5-804a-c69eba691629 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.282029] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ec4ddf4-43b0-46e6-97f2-e0e4f2533d43 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.287793] env[63021]: DEBUG nova.compute.manager [req-8b3ebd26-a931-4b82-a364-85bc4f090888 req-594eaf1f-2f4a-4fe5-b37c-e12e8a8725d9 service nova] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Received event network-vif-plugged-ef03a045-2fd4-4794-964b-ec6a72e8f7d2 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1067.288016] env[63021]: DEBUG oslo_concurrency.lockutils [req-8b3ebd26-a931-4b82-a364-85bc4f090888 req-594eaf1f-2f4a-4fe5-b37c-e12e8a8725d9 service nova] Acquiring lock "a2d6db04-9293-4755-ad8c-2e46467f92cb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.288273] env[63021]: DEBUG oslo_concurrency.lockutils [req-8b3ebd26-a931-4b82-a364-85bc4f090888 req-594eaf1f-2f4a-4fe5-b37c-e12e8a8725d9 service nova] Lock "a2d6db04-9293-4755-ad8c-2e46467f92cb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.288455] env[63021]: DEBUG oslo_concurrency.lockutils [req-8b3ebd26-a931-4b82-a364-85bc4f090888 req-594eaf1f-2f4a-4fe5-b37c-e12e8a8725d9 service nova] Lock "a2d6db04-9293-4755-ad8c-2e46467f92cb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.288625] env[63021]: DEBUG nova.compute.manager [req-8b3ebd26-a931-4b82-a364-85bc4f090888 req-594eaf1f-2f4a-4fe5-b37c-e12e8a8725d9 service nova] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] No waiting events found dispatching network-vif-plugged-ef03a045-2fd4-4794-964b-ec6a72e8f7d2 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1067.288790] env[63021]: WARNING nova.compute.manager [req-8b3ebd26-a931-4b82-a364-85bc4f090888 req-594eaf1f-2f4a-4fe5-b37c-e12e8a8725d9 service nova] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Received unexpected event network-vif-plugged-ef03a045-2fd4-4794-964b-ec6a72e8f7d2 for instance with vm_state building and task_state spawning. [ 1067.298645] env[63021]: DEBUG nova.compute.provider_tree [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1067.361328] env[63021]: DEBUG nova.compute.manager [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1067.388885] env[63021]: DEBUG nova.network.neutron [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Successfully updated port: ef03a045-2fd4-4794-964b-ec6a72e8f7d2 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1067.801519] env[63021]: DEBUG nova.scheduler.client.report [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1067.884154] env[63021]: DEBUG oslo_concurrency.lockutils [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.891037] env[63021]: DEBUG oslo_concurrency.lockutils [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "refresh_cache-a2d6db04-9293-4755-ad8c-2e46467f92cb" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.891146] env[63021]: DEBUG oslo_concurrency.lockutils [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquired lock "refresh_cache-a2d6db04-9293-4755-ad8c-2e46467f92cb" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.891229] env[63021]: DEBUG nova.network.neutron [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1068.306343] env[63021]: DEBUG oslo_concurrency.lockutils [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.228s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.306689] env[63021]: INFO nova.compute.manager [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Migrating [ 1068.313451] env[63021]: DEBUG oslo_concurrency.lockutils [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.430s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.314906] env[63021]: INFO nova.compute.claims [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1068.430413] env[63021]: DEBUG nova.network.neutron [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1068.554696] env[63021]: DEBUG nova.network.neutron [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Updating instance_info_cache with network_info: [{"id": "ef03a045-2fd4-4794-964b-ec6a72e8f7d2", "address": "fa:16:3e:16:2d:18", "network": {"id": "0c0dd1a7-6687-45c5-82b1-33f34bdbb848", "bridge": "br-int", "label": "tempest-ServersTestJSON-523240467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c8959d2f1fe4706b7407c11e128b773", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffc811e4-0e4e-4748-8c7e-b3f14ccbd42d", "external-id": "nsx-vlan-transportzone-404", "segmentation_id": 404, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef03a045-2f", "ovs_interfaceid": "ef03a045-2fd4-4794-964b-ec6a72e8f7d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.827157] env[63021]: DEBUG oslo_concurrency.lockutils [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "refresh_cache-e940f625-f35e-459d-b8be-fa5916ce6efb" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1068.827370] env[63021]: DEBUG oslo_concurrency.lockutils [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquired lock "refresh_cache-e940f625-f35e-459d-b8be-fa5916ce6efb" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.827554] env[63021]: DEBUG nova.network.neutron [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1069.057143] env[63021]: DEBUG oslo_concurrency.lockutils [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Releasing lock "refresh_cache-a2d6db04-9293-4755-ad8c-2e46467f92cb" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1069.057722] env[63021]: DEBUG nova.compute.manager [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Instance network_info: |[{"id": "ef03a045-2fd4-4794-964b-ec6a72e8f7d2", "address": "fa:16:3e:16:2d:18", "network": {"id": "0c0dd1a7-6687-45c5-82b1-33f34bdbb848", "bridge": "br-int", "label": "tempest-ServersTestJSON-523240467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c8959d2f1fe4706b7407c11e128b773", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffc811e4-0e4e-4748-8c7e-b3f14ccbd42d", "external-id": "nsx-vlan-transportzone-404", "segmentation_id": 404, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef03a045-2f", "ovs_interfaceid": "ef03a045-2fd4-4794-964b-ec6a72e8f7d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1069.058525] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:16:2d:18', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ffc811e4-0e4e-4748-8c7e-b3f14ccbd42d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ef03a045-2fd4-4794-964b-ec6a72e8f7d2', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1069.069818] env[63021]: DEBUG oslo.service.loopingcall [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1069.070137] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1069.070432] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-30716dce-fbde-4cf6-bc7a-3ef6fea75c12 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.096110] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1069.096110] env[63021]: value = "task-1294273" [ 1069.096110] env[63021]: _type = "Task" [ 1069.096110] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.103998] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294273, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.317886] env[63021]: DEBUG nova.compute.manager [req-43f1628a-6550-433a-a33e-ad78ea5bc775 req-fd0d45b5-3dee-429b-a7af-f3dc1b87702d service nova] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Received event network-changed-ef03a045-2fd4-4794-964b-ec6a72e8f7d2 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1069.318152] env[63021]: DEBUG nova.compute.manager [req-43f1628a-6550-433a-a33e-ad78ea5bc775 req-fd0d45b5-3dee-429b-a7af-f3dc1b87702d service nova] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Refreshing instance network info cache due to event network-changed-ef03a045-2fd4-4794-964b-ec6a72e8f7d2. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1069.318316] env[63021]: DEBUG oslo_concurrency.lockutils [req-43f1628a-6550-433a-a33e-ad78ea5bc775 req-fd0d45b5-3dee-429b-a7af-f3dc1b87702d service nova] Acquiring lock "refresh_cache-a2d6db04-9293-4755-ad8c-2e46467f92cb" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.318477] env[63021]: DEBUG oslo_concurrency.lockutils [req-43f1628a-6550-433a-a33e-ad78ea5bc775 req-fd0d45b5-3dee-429b-a7af-f3dc1b87702d service nova] Acquired lock "refresh_cache-a2d6db04-9293-4755-ad8c-2e46467f92cb" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.318644] env[63021]: DEBUG nova.network.neutron [req-43f1628a-6550-433a-a33e-ad78ea5bc775 req-fd0d45b5-3dee-429b-a7af-f3dc1b87702d service nova] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Refreshing network info cache for port ef03a045-2fd4-4794-964b-ec6a72e8f7d2 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1069.479253] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31d6f452-06c2-4dda-9151-e1f8ce37799b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.486921] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b128da57-6d4c-4c29-a22b-20cc5c6db4f0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.520751] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-833ea92c-8815-4e89-83dc-0da20a160c49 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.527986] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57b52350-c6d5-4ed6-8d36-eb305b9eb167 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.540673] env[63021]: DEBUG nova.compute.provider_tree [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1069.591191] env[63021]: DEBUG nova.network.neutron [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Updating instance_info_cache with network_info: [{"id": "d2876adc-ed3c-4540-b06d-fd179ecd2445", "address": "fa:16:3e:a2:13:81", "network": {"id": "81917a55-4e41-4898-9f4e-4fced52b5a86", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1793837736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e238d5c1668540d2a9fffd3fd832d9b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2876adc-ed", "ovs_interfaceid": "d2876adc-ed3c-4540-b06d-fd179ecd2445", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.605948] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294273, 'name': CreateVM_Task, 'duration_secs': 0.331434} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.606126] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1069.606770] env[63021]: DEBUG oslo_concurrency.lockutils [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.606939] env[63021]: DEBUG oslo_concurrency.lockutils [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.607440] env[63021]: DEBUG oslo_concurrency.lockutils [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1069.607685] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5201492-bbb0-4893-8592-22ad818244f0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.612078] env[63021]: DEBUG oslo_vmware.api [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1069.612078] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]524455b0-c8cd-dfff-1d46-631c5d934708" [ 1069.612078] env[63021]: _type = "Task" [ 1069.612078] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.619526] env[63021]: DEBUG oslo_vmware.api [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]524455b0-c8cd-dfff-1d46-631c5d934708, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.791325] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-3db7651d-23fa-4bc2-9db6-725c3ef3c5b3 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Volume attach. Driver type: vmdk {{(pid=63021) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1069.791598] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-3db7651d-23fa-4bc2-9db6-725c3ef3c5b3 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277595', 'volume_id': '29440f30-19f1-42a5-9503-a29373b566ff', 'name': 'volume-29440f30-19f1-42a5-9503-a29373b566ff', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8d9387b8-6a84-4ada-8ac7-4ffd922b675a', 'attached_at': '', 'detached_at': '', 'volume_id': '29440f30-19f1-42a5-9503-a29373b566ff', 'serial': '29440f30-19f1-42a5-9503-a29373b566ff'} {{(pid=63021) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1069.793034] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9032cfa-84af-4eb6-85ba-b4c7608e32d8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.808733] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fc4e457-8300-4024-bd47-db3b242ab2ff {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.834215] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-3db7651d-23fa-4bc2-9db6-725c3ef3c5b3 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] volume-29440f30-19f1-42a5-9503-a29373b566ff/volume-29440f30-19f1-42a5-9503-a29373b566ff.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1069.834510] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c9ab1fe-8182-458c-b739-d40389e16e37 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.852717] env[63021]: DEBUG oslo_vmware.api [None req-3db7651d-23fa-4bc2-9db6-725c3ef3c5b3 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 1069.852717] env[63021]: value = "task-1294274" [ 1069.852717] env[63021]: _type = "Task" [ 1069.852717] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.862819] env[63021]: DEBUG oslo_vmware.api [None req-3db7651d-23fa-4bc2-9db6-725c3ef3c5b3 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294274, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.043362] env[63021]: DEBUG nova.scheduler.client.report [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1070.093848] env[63021]: DEBUG oslo_concurrency.lockutils [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Releasing lock "refresh_cache-e940f625-f35e-459d-b8be-fa5916ce6efb" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.127215] env[63021]: DEBUG oslo_vmware.api [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]524455b0-c8cd-dfff-1d46-631c5d934708, 'name': SearchDatastore_Task, 'duration_secs': 0.00959} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.128469] env[63021]: DEBUG nova.network.neutron [req-43f1628a-6550-433a-a33e-ad78ea5bc775 req-fd0d45b5-3dee-429b-a7af-f3dc1b87702d service nova] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Updated VIF entry in instance network info cache for port ef03a045-2fd4-4794-964b-ec6a72e8f7d2. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1070.128819] env[63021]: DEBUG nova.network.neutron [req-43f1628a-6550-433a-a33e-ad78ea5bc775 req-fd0d45b5-3dee-429b-a7af-f3dc1b87702d service nova] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Updating instance_info_cache with network_info: [{"id": "ef03a045-2fd4-4794-964b-ec6a72e8f7d2", "address": "fa:16:3e:16:2d:18", "network": {"id": "0c0dd1a7-6687-45c5-82b1-33f34bdbb848", "bridge": "br-int", "label": "tempest-ServersTestJSON-523240467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c8959d2f1fe4706b7407c11e128b773", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffc811e4-0e4e-4748-8c7e-b3f14ccbd42d", "external-id": "nsx-vlan-transportzone-404", "segmentation_id": 404, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef03a045-2f", "ovs_interfaceid": "ef03a045-2fd4-4794-964b-ec6a72e8f7d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.130154] env[63021]: DEBUG oslo_concurrency.lockutils [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.130321] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1070.130592] env[63021]: DEBUG oslo_concurrency.lockutils [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.130856] env[63021]: DEBUG oslo_concurrency.lockutils [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.131024] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1070.131603] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-00ce67cb-7cd1-4286-b7ed-77812bc0ac99 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.140649] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1070.140772] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1070.141626] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca4a7a54-0748-42ad-9c32-4cb9ee6bb057 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.147345] env[63021]: DEBUG oslo_vmware.api [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1070.147345] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5229c076-0724-0d7a-2660-8a69049a5602" [ 1070.147345] env[63021]: _type = "Task" [ 1070.147345] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.155838] env[63021]: DEBUG oslo_vmware.api [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5229c076-0724-0d7a-2660-8a69049a5602, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.234758] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e571f476-5708-4f52-9bfd-679d05b0fca6 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "d03c47da-7e73-4474-ad9b-07d27629dcae" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.235018] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e571f476-5708-4f52-9bfd-679d05b0fca6 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "d03c47da-7e73-4474-ad9b-07d27629dcae" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.362948] env[63021]: DEBUG oslo_vmware.api [None req-3db7651d-23fa-4bc2-9db6-725c3ef3c5b3 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294274, 'name': ReconfigVM_Task, 'duration_secs': 0.359376} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.363267] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-3db7651d-23fa-4bc2-9db6-725c3ef3c5b3 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Reconfigured VM instance instance-00000061 to attach disk [datastore1] volume-29440f30-19f1-42a5-9503-a29373b566ff/volume-29440f30-19f1-42a5-9503-a29373b566ff.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1070.368142] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4c527f62-4abc-4965-919e-f305e43496de {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.383215] env[63021]: DEBUG oslo_vmware.api [None req-3db7651d-23fa-4bc2-9db6-725c3ef3c5b3 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 1070.383215] env[63021]: value = "task-1294275" [ 1070.383215] env[63021]: _type = "Task" [ 1070.383215] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.390511] env[63021]: DEBUG oslo_vmware.api [None req-3db7651d-23fa-4bc2-9db6-725c3ef3c5b3 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294275, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.548917] env[63021]: DEBUG oslo_concurrency.lockutils [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.235s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.549600] env[63021]: DEBUG nova.compute.manager [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1070.632285] env[63021]: DEBUG oslo_concurrency.lockutils [req-43f1628a-6550-433a-a33e-ad78ea5bc775 req-fd0d45b5-3dee-429b-a7af-f3dc1b87702d service nova] Releasing lock "refresh_cache-a2d6db04-9293-4755-ad8c-2e46467f92cb" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.658287] env[63021]: DEBUG oslo_vmware.api [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5229c076-0724-0d7a-2660-8a69049a5602, 'name': SearchDatastore_Task, 'duration_secs': 0.009016} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.659210] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee095cc7-c0e5-454b-8b32-028babc54936 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.664813] env[63021]: DEBUG oslo_vmware.api [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1070.664813] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]528834a8-895a-1227-d2f3-056dab808b48" [ 1070.664813] env[63021]: _type = "Task" [ 1070.664813] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.673131] env[63021]: DEBUG oslo_vmware.api [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]528834a8-895a-1227-d2f3-056dab808b48, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.738018] env[63021]: DEBUG nova.compute.utils [None req-e571f476-5708-4f52-9bfd-679d05b0fca6 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1070.893016] env[63021]: DEBUG oslo_vmware.api [None req-3db7651d-23fa-4bc2-9db6-725c3ef3c5b3 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294275, 'name': ReconfigVM_Task, 'duration_secs': 0.149587} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.893278] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-3db7651d-23fa-4bc2-9db6-725c3ef3c5b3 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277595', 'volume_id': '29440f30-19f1-42a5-9503-a29373b566ff', 'name': 'volume-29440f30-19f1-42a5-9503-a29373b566ff', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8d9387b8-6a84-4ada-8ac7-4ffd922b675a', 'attached_at': '', 'detached_at': '', 'volume_id': '29440f30-19f1-42a5-9503-a29373b566ff', 'serial': '29440f30-19f1-42a5-9503-a29373b566ff'} {{(pid=63021) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1071.054448] env[63021]: DEBUG nova.compute.utils [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1071.055859] env[63021]: DEBUG nova.compute.manager [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1071.056039] env[63021]: DEBUG nova.network.neutron [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1071.101655] env[63021]: DEBUG nova.policy [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '31c5d292015043aa8bfc6647e7405ea3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5af87713ae6b477ca8a584aed287889d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 1071.178765] env[63021]: DEBUG oslo_vmware.api [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]528834a8-895a-1227-d2f3-056dab808b48, 'name': SearchDatastore_Task, 'duration_secs': 0.009733} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.179578] env[63021]: DEBUG oslo_concurrency.lockutils [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.179951] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] a2d6db04-9293-4755-ad8c-2e46467f92cb/a2d6db04-9293-4755-ad8c-2e46467f92cb.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1071.180309] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dce99679-514b-41d6-83db-602d6669e4bb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.187543] env[63021]: DEBUG oslo_vmware.api [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1071.187543] env[63021]: value = "task-1294276" [ 1071.187543] env[63021]: _type = "Task" [ 1071.187543] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.197694] env[63021]: DEBUG oslo_vmware.api [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294276, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.240688] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e571f476-5708-4f52-9bfd-679d05b0fca6 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "d03c47da-7e73-4474-ad9b-07d27629dcae" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.454831] env[63021]: DEBUG nova.network.neutron [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Successfully created port: b9d1b591-e9bc-4e70-9c76-1a0673bca1d5 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1071.559679] env[63021]: DEBUG nova.compute.manager [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1071.613072] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dd4ebe6-b122-4e8f-aeb5-d90546328fc2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.635832] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Updating instance 'e940f625-f35e-459d-b8be-fa5916ce6efb' progress to 0 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1071.700789] env[63021]: DEBUG oslo_vmware.api [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294276, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.939167] env[63021]: DEBUG nova.objects.instance [None req-3db7651d-23fa-4bc2-9db6-725c3ef3c5b3 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lazy-loading 'flavor' on Instance uuid 8d9387b8-6a84-4ada-8ac7-4ffd922b675a {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1072.142573] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1072.142900] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4610c96b-ba89-42b8-9a6c-2f76f2a95064 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.150515] env[63021]: DEBUG oslo_vmware.api [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1072.150515] env[63021]: value = "task-1294277" [ 1072.150515] env[63021]: _type = "Task" [ 1072.150515] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.159146] env[63021]: DEBUG oslo_vmware.api [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294277, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.200593] env[63021]: DEBUG oslo_vmware.api [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294276, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.577068} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.200902] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] a2d6db04-9293-4755-ad8c-2e46467f92cb/a2d6db04-9293-4755-ad8c-2e46467f92cb.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1072.201139] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1072.201402] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-87267b28-cc88-4aef-b9fb-30135c29e51d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.208183] env[63021]: DEBUG oslo_vmware.api [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1072.208183] env[63021]: value = "task-1294278" [ 1072.208183] env[63021]: _type = "Task" [ 1072.208183] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.216537] env[63021]: DEBUG oslo_vmware.api [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294278, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.307043] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e571f476-5708-4f52-9bfd-679d05b0fca6 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "d03c47da-7e73-4474-ad9b-07d27629dcae" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.307352] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e571f476-5708-4f52-9bfd-679d05b0fca6 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "d03c47da-7e73-4474-ad9b-07d27629dcae" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.307598] env[63021]: INFO nova.compute.manager [None req-e571f476-5708-4f52-9bfd-679d05b0fca6 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Attaching volume cd524289-5d08-4b7f-a083-259e501f1fc1 to /dev/sdb [ 1072.341037] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb6ab50c-8b5c-423c-bfe2-685430528618 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.348842] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec5dafe7-20da-4688-845e-c550ba9b015a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.362893] env[63021]: DEBUG nova.virt.block_device [None req-e571f476-5708-4f52-9bfd-679d05b0fca6 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Updating existing volume attachment record: 6940a357-3401-4fa9-9fa0-5cb651332046 {{(pid=63021) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1072.444805] env[63021]: DEBUG oslo_concurrency.lockutils [None req-3db7651d-23fa-4bc2-9db6-725c3ef3c5b3 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "8d9387b8-6a84-4ada-8ac7-4ffd922b675a" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.252s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.570296] env[63021]: DEBUG nova.compute.manager [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1072.594990] env[63021]: DEBUG nova.virt.hardware [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1072.595280] env[63021]: DEBUG nova.virt.hardware [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1072.595442] env[63021]: DEBUG nova.virt.hardware [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1072.595646] env[63021]: DEBUG nova.virt.hardware [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1072.595797] env[63021]: DEBUG nova.virt.hardware [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1072.595948] env[63021]: DEBUG nova.virt.hardware [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1072.596218] env[63021]: DEBUG nova.virt.hardware [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1072.596426] env[63021]: DEBUG nova.virt.hardware [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1072.596607] env[63021]: DEBUG nova.virt.hardware [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1072.596774] env[63021]: DEBUG nova.virt.hardware [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1072.596947] env[63021]: DEBUG nova.virt.hardware [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1072.597864] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dec704b5-d761-499a-bb16-aff1d299f1cb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.605568] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39129980-54a3-41f0-973d-293009159c23 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.659593] env[63021]: DEBUG oslo_vmware.api [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294277, 'name': PowerOffVM_Task, 'duration_secs': 0.282564} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.659821] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1072.660010] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Updating instance 'e940f625-f35e-459d-b8be-fa5916ce6efb' progress to 17 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1072.718657] env[63021]: DEBUG oslo_vmware.api [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294278, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066515} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.718934] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1072.719717] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7961c69b-b31c-41e8-abc6-ddbbdd0e2669 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.741642] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] a2d6db04-9293-4755-ad8c-2e46467f92cb/a2d6db04-9293-4755-ad8c-2e46467f92cb.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1072.742248] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f0389d17-f216-4d5f-aab3-8a417e8398aa {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.760994] env[63021]: DEBUG oslo_vmware.api [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1072.760994] env[63021]: value = "task-1294282" [ 1072.760994] env[63021]: _type = "Task" [ 1072.760994] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.768521] env[63021]: DEBUG oslo_vmware.api [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294282, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.795629] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1a64bdb8-805e-46aa-9dd9-6582c1829258 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "67f736d9-82ec-448e-8b72-84fdd704aa8d" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.795774] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1a64bdb8-805e-46aa-9dd9-6582c1829258 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "67f736d9-82ec-448e-8b72-84fdd704aa8d" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.038020] env[63021]: DEBUG nova.compute.manager [req-bb547afb-3386-4af7-bdde-97c25bc36dfe req-65d78ed8-8d02-4ef3-9280-023e931ae267 service nova] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Received event network-vif-plugged-b9d1b591-e9bc-4e70-9c76-1a0673bca1d5 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1073.038314] env[63021]: DEBUG oslo_concurrency.lockutils [req-bb547afb-3386-4af7-bdde-97c25bc36dfe req-65d78ed8-8d02-4ef3-9280-023e931ae267 service nova] Acquiring lock "4041dfbf-fa94-4ff8-9176-e7088cdf241f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.038541] env[63021]: DEBUG oslo_concurrency.lockutils [req-bb547afb-3386-4af7-bdde-97c25bc36dfe req-65d78ed8-8d02-4ef3-9280-023e931ae267 service nova] Lock "4041dfbf-fa94-4ff8-9176-e7088cdf241f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.038755] env[63021]: DEBUG oslo_concurrency.lockutils [req-bb547afb-3386-4af7-bdde-97c25bc36dfe req-65d78ed8-8d02-4ef3-9280-023e931ae267 service nova] Lock "4041dfbf-fa94-4ff8-9176-e7088cdf241f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.038945] env[63021]: DEBUG nova.compute.manager [req-bb547afb-3386-4af7-bdde-97c25bc36dfe req-65d78ed8-8d02-4ef3-9280-023e931ae267 service nova] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] No waiting events found dispatching network-vif-plugged-b9d1b591-e9bc-4e70-9c76-1a0673bca1d5 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1073.039172] env[63021]: WARNING nova.compute.manager [req-bb547afb-3386-4af7-bdde-97c25bc36dfe req-65d78ed8-8d02-4ef3-9280-023e931ae267 service nova] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Received unexpected event network-vif-plugged-b9d1b591-e9bc-4e70-9c76-1a0673bca1d5 for instance with vm_state building and task_state spawning. [ 1073.126501] env[63021]: DEBUG nova.network.neutron [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Successfully updated port: b9d1b591-e9bc-4e70-9c76-1a0673bca1d5 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1073.166598] env[63021]: DEBUG nova.virt.hardware [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1073.166885] env[63021]: DEBUG nova.virt.hardware [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1073.167057] env[63021]: DEBUG nova.virt.hardware [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1073.167286] env[63021]: DEBUG nova.virt.hardware [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1073.167437] env[63021]: DEBUG nova.virt.hardware [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1073.167587] env[63021]: DEBUG nova.virt.hardware [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1073.167794] env[63021]: DEBUG nova.virt.hardware [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1073.167951] env[63021]: DEBUG nova.virt.hardware [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1073.168133] env[63021]: DEBUG nova.virt.hardware [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1073.168300] env[63021]: DEBUG nova.virt.hardware [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1073.168474] env[63021]: DEBUG nova.virt.hardware [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1073.173834] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-14b8d94f-8e86-4333-a9c5-5d642b35b269 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.189015] env[63021]: DEBUG oslo_vmware.api [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1073.189015] env[63021]: value = "task-1294283" [ 1073.189015] env[63021]: _type = "Task" [ 1073.189015] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.197009] env[63021]: DEBUG oslo_vmware.api [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294283, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.271446] env[63021]: DEBUG oslo_vmware.api [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294282, 'name': ReconfigVM_Task, 'duration_secs': 0.266633} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.271766] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Reconfigured VM instance instance-00000068 to attach disk [datastore1] a2d6db04-9293-4755-ad8c-2e46467f92cb/a2d6db04-9293-4755-ad8c-2e46467f92cb.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1073.272407] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aab92ec7-f66c-4e99-b4b7-5fd2e666e182 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.278714] env[63021]: DEBUG oslo_vmware.api [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1073.278714] env[63021]: value = "task-1294284" [ 1073.278714] env[63021]: _type = "Task" [ 1073.278714] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.286643] env[63021]: DEBUG oslo_vmware.api [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294284, 'name': Rename_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.297827] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1f365d8b-2d16-46bd-af71-e14f2b9a311c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquiring lock "8d9387b8-6a84-4ada-8ac7-4ffd922b675a" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.298071] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1f365d8b-2d16-46bd-af71-e14f2b9a311c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "8d9387b8-6a84-4ada-8ac7-4ffd922b675a" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.299907] env[63021]: INFO nova.compute.manager [None req-1a64bdb8-805e-46aa-9dd9-6582c1829258 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Detaching volume 236a8786-fa3f-411d-a486-ddd5f2950147 [ 1073.332674] env[63021]: INFO nova.virt.block_device [None req-1a64bdb8-805e-46aa-9dd9-6582c1829258 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Attempting to driver detach volume 236a8786-fa3f-411d-a486-ddd5f2950147 from mountpoint /dev/sdb [ 1073.333139] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a64bdb8-805e-46aa-9dd9-6582c1829258 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Volume detach. Driver type: vmdk {{(pid=63021) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1073.333212] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a64bdb8-805e-46aa-9dd9-6582c1829258 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277584', 'volume_id': '236a8786-fa3f-411d-a486-ddd5f2950147', 'name': 'volume-236a8786-fa3f-411d-a486-ddd5f2950147', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '67f736d9-82ec-448e-8b72-84fdd704aa8d', 'attached_at': '', 'detached_at': '', 'volume_id': '236a8786-fa3f-411d-a486-ddd5f2950147', 'serial': '236a8786-fa3f-411d-a486-ddd5f2950147'} {{(pid=63021) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1073.334443] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-336de669-97e0-4a24-ad37-6dfa33705470 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.356547] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27b75c54-04a6-47f7-8361-4268af7fc206 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.363499] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ee48ab1-02a3-4fc5-98d4-db1641c43920 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.385188] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e63be0f-12b5-45fe-a4cf-a38fe0f94f98 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.402332] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a64bdb8-805e-46aa-9dd9-6582c1829258 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] The volume has not been displaced from its original location: [datastore2] volume-236a8786-fa3f-411d-a486-ddd5f2950147/volume-236a8786-fa3f-411d-a486-ddd5f2950147.vmdk. No consolidation needed. {{(pid=63021) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1073.407735] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a64bdb8-805e-46aa-9dd9-6582c1829258 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Reconfiguring VM instance instance-0000005b to detach disk 2001 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1073.408065] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-08392ed3-a54d-4501-83a0-d19d916d5867 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.425368] env[63021]: DEBUG oslo_vmware.api [None req-1a64bdb8-805e-46aa-9dd9-6582c1829258 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 1073.425368] env[63021]: value = "task-1294285" [ 1073.425368] env[63021]: _type = "Task" [ 1073.425368] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.433684] env[63021]: DEBUG oslo_vmware.api [None req-1a64bdb8-805e-46aa-9dd9-6582c1829258 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294285, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.629086] env[63021]: DEBUG oslo_concurrency.lockutils [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "refresh_cache-4041dfbf-fa94-4ff8-9176-e7088cdf241f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.629313] env[63021]: DEBUG oslo_concurrency.lockutils [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquired lock "refresh_cache-4041dfbf-fa94-4ff8-9176-e7088cdf241f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.629493] env[63021]: DEBUG nova.network.neutron [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1073.699019] env[63021]: DEBUG oslo_vmware.api [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294283, 'name': ReconfigVM_Task, 'duration_secs': 0.230005} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.699334] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Updating instance 'e940f625-f35e-459d-b8be-fa5916ce6efb' progress to 33 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1073.789312] env[63021]: DEBUG oslo_vmware.api [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294284, 'name': Rename_Task, 'duration_secs': 0.14634} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.789611] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1073.789853] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2e4deace-3661-421c-b156-8fc9b1715ef6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.796473] env[63021]: DEBUG oslo_vmware.api [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1073.796473] env[63021]: value = "task-1294286" [ 1073.796473] env[63021]: _type = "Task" [ 1073.796473] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.804750] env[63021]: DEBUG nova.compute.utils [None req-1f365d8b-2d16-46bd-af71-e14f2b9a311c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1073.805947] env[63021]: DEBUG oslo_vmware.api [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294286, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.935140] env[63021]: DEBUG oslo_vmware.api [None req-1a64bdb8-805e-46aa-9dd9-6582c1829258 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294285, 'name': ReconfigVM_Task, 'duration_secs': 0.214906} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.935482] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a64bdb8-805e-46aa-9dd9-6582c1829258 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Reconfigured VM instance instance-0000005b to detach disk 2001 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1073.941496] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0d712945-e2f4-4206-9bd6-9dacf57ee193 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.961332] env[63021]: DEBUG oslo_vmware.api [None req-1a64bdb8-805e-46aa-9dd9-6582c1829258 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 1073.961332] env[63021]: value = "task-1294287" [ 1073.961332] env[63021]: _type = "Task" [ 1073.961332] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.970303] env[63021]: DEBUG oslo_vmware.api [None req-1a64bdb8-805e-46aa-9dd9-6582c1829258 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294287, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.161209] env[63021]: DEBUG nova.network.neutron [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1074.205533] env[63021]: DEBUG nova.virt.hardware [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1074.205796] env[63021]: DEBUG nova.virt.hardware [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1074.205959] env[63021]: DEBUG nova.virt.hardware [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1074.206161] env[63021]: DEBUG nova.virt.hardware [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1074.206313] env[63021]: DEBUG nova.virt.hardware [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1074.206462] env[63021]: DEBUG nova.virt.hardware [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1074.206675] env[63021]: DEBUG nova.virt.hardware [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1074.206845] env[63021]: DEBUG nova.virt.hardware [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1074.207067] env[63021]: DEBUG nova.virt.hardware [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1074.207251] env[63021]: DEBUG nova.virt.hardware [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1074.207445] env[63021]: DEBUG nova.virt.hardware [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1074.212816] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Reconfiguring VM instance instance-00000067 to detach disk 2000 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1074.215380] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ca6684a8-6d9c-4fbb-9778-e2cb935d0921 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.234521] env[63021]: DEBUG oslo_vmware.api [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1074.234521] env[63021]: value = "task-1294288" [ 1074.234521] env[63021]: _type = "Task" [ 1074.234521] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.243179] env[63021]: DEBUG oslo_vmware.api [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294288, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.309242] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1f365d8b-2d16-46bd-af71-e14f2b9a311c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "8d9387b8-6a84-4ada-8ac7-4ffd922b675a" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.011s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.309818] env[63021]: DEBUG oslo_vmware.api [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294286, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.337546] env[63021]: DEBUG nova.network.neutron [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Updating instance_info_cache with network_info: [{"id": "b9d1b591-e9bc-4e70-9c76-1a0673bca1d5", "address": "fa:16:3e:1b:b0:e6", "network": {"id": "9502e364-60e1-47ef-a1f7-f3e3ff45cf69", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-164717784-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5af87713ae6b477ca8a584aed287889d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9d1b591-e9", "ovs_interfaceid": "b9d1b591-e9bc-4e70-9c76-1a0673bca1d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1074.470970] env[63021]: DEBUG oslo_vmware.api [None req-1a64bdb8-805e-46aa-9dd9-6582c1829258 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294287, 'name': ReconfigVM_Task, 'duration_secs': 0.155048} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.471304] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a64bdb8-805e-46aa-9dd9-6582c1829258 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277584', 'volume_id': '236a8786-fa3f-411d-a486-ddd5f2950147', 'name': 'volume-236a8786-fa3f-411d-a486-ddd5f2950147', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '67f736d9-82ec-448e-8b72-84fdd704aa8d', 'attached_at': '', 'detached_at': '', 'volume_id': '236a8786-fa3f-411d-a486-ddd5f2950147', 'serial': '236a8786-fa3f-411d-a486-ddd5f2950147'} {{(pid=63021) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1074.744442] env[63021]: DEBUG oslo_vmware.api [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294288, 'name': ReconfigVM_Task, 'duration_secs': 0.213854} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.744743] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Reconfigured VM instance instance-00000067 to detach disk 2000 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1074.745499] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d40758a-6b18-460b-8e35-0e6b880f2da0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.767998] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] e940f625-f35e-459d-b8be-fa5916ce6efb/e940f625-f35e-459d-b8be-fa5916ce6efb.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1074.768314] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-519327d5-710b-4d03-919c-ab63ccf1638a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.785498] env[63021]: DEBUG oslo_vmware.api [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1074.785498] env[63021]: value = "task-1294290" [ 1074.785498] env[63021]: _type = "Task" [ 1074.785498] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.793191] env[63021]: DEBUG oslo_vmware.api [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294290, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.804954] env[63021]: DEBUG oslo_vmware.api [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294286, 'name': PowerOnVM_Task, 'duration_secs': 0.515111} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.805357] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1074.805665] env[63021]: INFO nova.compute.manager [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Took 7.81 seconds to spawn the instance on the hypervisor. [ 1074.805952] env[63021]: DEBUG nova.compute.manager [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1074.807161] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5937793a-4504-4384-839b-7a2a2df29fa4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.840809] env[63021]: DEBUG oslo_concurrency.lockutils [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Releasing lock "refresh_cache-4041dfbf-fa94-4ff8-9176-e7088cdf241f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.841165] env[63021]: DEBUG nova.compute.manager [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Instance network_info: |[{"id": "b9d1b591-e9bc-4e70-9c76-1a0673bca1d5", "address": "fa:16:3e:1b:b0:e6", "network": {"id": "9502e364-60e1-47ef-a1f7-f3e3ff45cf69", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-164717784-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5af87713ae6b477ca8a584aed287889d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9d1b591-e9", "ovs_interfaceid": "b9d1b591-e9bc-4e70-9c76-1a0673bca1d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1074.841694] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1b:b0:e6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b9d1b591-e9bc-4e70-9c76-1a0673bca1d5', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1074.849417] env[63021]: DEBUG oslo.service.loopingcall [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1074.850403] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1074.850644] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bf913a81-0425-413c-98b2-f133a6fc4231 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.869875] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1074.869875] env[63021]: value = "task-1294291" [ 1074.869875] env[63021]: _type = "Task" [ 1074.869875] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.878040] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294291, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.013845] env[63021]: DEBUG nova.objects.instance [None req-1a64bdb8-805e-46aa-9dd9-6582c1829258 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lazy-loading 'flavor' on Instance uuid 67f736d9-82ec-448e-8b72-84fdd704aa8d {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1075.111852] env[63021]: DEBUG nova.compute.manager [req-b6f5af7f-eeff-4df2-aa1b-6949599fea37 req-fa3f40c8-da31-4e88-bc96-26223b8dc177 service nova] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Received event network-changed-b9d1b591-e9bc-4e70-9c76-1a0673bca1d5 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1075.112216] env[63021]: DEBUG nova.compute.manager [req-b6f5af7f-eeff-4df2-aa1b-6949599fea37 req-fa3f40c8-da31-4e88-bc96-26223b8dc177 service nova] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Refreshing instance network info cache due to event network-changed-b9d1b591-e9bc-4e70-9c76-1a0673bca1d5. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1075.112302] env[63021]: DEBUG oslo_concurrency.lockutils [req-b6f5af7f-eeff-4df2-aa1b-6949599fea37 req-fa3f40c8-da31-4e88-bc96-26223b8dc177 service nova] Acquiring lock "refresh_cache-4041dfbf-fa94-4ff8-9176-e7088cdf241f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1075.112474] env[63021]: DEBUG oslo_concurrency.lockutils [req-b6f5af7f-eeff-4df2-aa1b-6949599fea37 req-fa3f40c8-da31-4e88-bc96-26223b8dc177 service nova] Acquired lock "refresh_cache-4041dfbf-fa94-4ff8-9176-e7088cdf241f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.112594] env[63021]: DEBUG nova.network.neutron [req-b6f5af7f-eeff-4df2-aa1b-6949599fea37 req-fa3f40c8-da31-4e88-bc96-26223b8dc177 service nova] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Refreshing network info cache for port b9d1b591-e9bc-4e70-9c76-1a0673bca1d5 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1075.295871] env[63021]: DEBUG oslo_vmware.api [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294290, 'name': ReconfigVM_Task, 'duration_secs': 0.296447} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.296421] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Reconfigured VM instance instance-00000067 to attach disk [datastore1] e940f625-f35e-459d-b8be-fa5916ce6efb/e940f625-f35e-459d-b8be-fa5916ce6efb.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1075.296736] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Updating instance 'e940f625-f35e-459d-b8be-fa5916ce6efb' progress to 50 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1075.325000] env[63021]: INFO nova.compute.manager [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Took 12.57 seconds to build instance. [ 1075.380563] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294291, 'name': CreateVM_Task, 'duration_secs': 0.316732} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.380701] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1075.381517] env[63021]: DEBUG oslo_concurrency.lockutils [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1075.381827] env[63021]: DEBUG oslo_concurrency.lockutils [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.382200] env[63021]: DEBUG oslo_concurrency.lockutils [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1075.382475] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b334e79-dbd7-4a01-9a44-12d46fedfe57 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.386931] env[63021]: DEBUG oslo_vmware.api [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1075.386931] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52091a21-247f-f010-54db-9b9b6a89fe0b" [ 1075.386931] env[63021]: _type = "Task" [ 1075.386931] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.394798] env[63021]: DEBUG oslo_vmware.api [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52091a21-247f-f010-54db-9b9b6a89fe0b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.594418] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1f365d8b-2d16-46bd-af71-e14f2b9a311c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquiring lock "8d9387b8-6a84-4ada-8ac7-4ffd922b675a" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.594729] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1f365d8b-2d16-46bd-af71-e14f2b9a311c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "8d9387b8-6a84-4ada-8ac7-4ffd922b675a" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.595445] env[63021]: INFO nova.compute.manager [None req-1f365d8b-2d16-46bd-af71-e14f2b9a311c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Attaching volume df640a1b-6005-45df-9ce3-d8634fcb4380 to /dev/sdc [ 1075.629113] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb9befe3-2f9c-4f60-8687-42dcb99fcf4e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.636557] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de1a25a5-26d9-430a-ae51-f6f0f54d186a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.651930] env[63021]: DEBUG nova.virt.block_device [None req-1f365d8b-2d16-46bd-af71-e14f2b9a311c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Updating existing volume attachment record: 61021355-f04c-4a7f-a3a8-1cb558d14bc0 {{(pid=63021) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1075.803258] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eba407d7-569f-497e-a6e5-a5a79cb96c0b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.824970] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a3e418f-9a97-4dd2-9a3f-4cb5346ccac5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.827729] env[63021]: DEBUG oslo_concurrency.lockutils [None req-173676eb-bb51-4b3e-ac97-ce4d41783489 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "a2d6db04-9293-4755-ad8c-2e46467f92cb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.083s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.843730] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Updating instance 'e940f625-f35e-459d-b8be-fa5916ce6efb' progress to 67 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1075.899120] env[63021]: DEBUG oslo_vmware.api [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52091a21-247f-f010-54db-9b9b6a89fe0b, 'name': SearchDatastore_Task, 'duration_secs': 0.010216} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.899359] env[63021]: DEBUG oslo_concurrency.lockutils [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1075.899604] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1075.899839] env[63021]: DEBUG oslo_concurrency.lockutils [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1075.899987] env[63021]: DEBUG oslo_concurrency.lockutils [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.900189] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1075.900463] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-95f632f6-4102-4cc8-ad2d-f7c9c5fd2daf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.909124] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1075.909309] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1075.910023] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e1175d3-20df-4cb6-8321-dacf12d91175 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.915568] env[63021]: DEBUG oslo_vmware.api [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1075.915568] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52da87b1-0c82-8b7d-7807-b22ef8b34fd9" [ 1075.915568] env[63021]: _type = "Task" [ 1075.915568] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.923251] env[63021]: DEBUG oslo_vmware.api [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52da87b1-0c82-8b7d-7807-b22ef8b34fd9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.926756] env[63021]: DEBUG nova.network.neutron [req-b6f5af7f-eeff-4df2-aa1b-6949599fea37 req-fa3f40c8-da31-4e88-bc96-26223b8dc177 service nova] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Updated VIF entry in instance network info cache for port b9d1b591-e9bc-4e70-9c76-1a0673bca1d5. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1075.927151] env[63021]: DEBUG nova.network.neutron [req-b6f5af7f-eeff-4df2-aa1b-6949599fea37 req-fa3f40c8-da31-4e88-bc96-26223b8dc177 service nova] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Updating instance_info_cache with network_info: [{"id": "b9d1b591-e9bc-4e70-9c76-1a0673bca1d5", "address": "fa:16:3e:1b:b0:e6", "network": {"id": "9502e364-60e1-47ef-a1f7-f3e3ff45cf69", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-164717784-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5af87713ae6b477ca8a584aed287889d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9d1b591-e9", "ovs_interfaceid": "b9d1b591-e9bc-4e70-9c76-1a0673bca1d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1076.020840] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1a64bdb8-805e-46aa-9dd9-6582c1829258 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "67f736d9-82ec-448e-8b72-84fdd704aa8d" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.225s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.383910] env[63021]: DEBUG nova.network.neutron [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Port d2876adc-ed3c-4540-b06d-fd179ecd2445 binding to destination host cpu-1 is already ACTIVE {{(pid=63021) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1076.425805] env[63021]: DEBUG oslo_vmware.api [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52da87b1-0c82-8b7d-7807-b22ef8b34fd9, 'name': SearchDatastore_Task, 'duration_secs': 0.010069} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.426759] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c139a227-88cd-4ed5-92b6-8d3b796d15f9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.429586] env[63021]: DEBUG oslo_concurrency.lockutils [req-b6f5af7f-eeff-4df2-aa1b-6949599fea37 req-fa3f40c8-da31-4e88-bc96-26223b8dc177 service nova] Releasing lock "refresh_cache-4041dfbf-fa94-4ff8-9176-e7088cdf241f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1076.433329] env[63021]: DEBUG oslo_vmware.api [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1076.433329] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52d17cf6-e18d-5191-139d-a4fa567dad6e" [ 1076.433329] env[63021]: _type = "Task" [ 1076.433329] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.442987] env[63021]: DEBUG oslo_vmware.api [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52d17cf6-e18d-5191-139d-a4fa567dad6e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.618657] env[63021]: DEBUG oslo_concurrency.lockutils [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "a2d6db04-9293-4755-ad8c-2e46467f92cb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.618971] env[63021]: DEBUG oslo_concurrency.lockutils [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "a2d6db04-9293-4755-ad8c-2e46467f92cb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.619207] env[63021]: DEBUG oslo_concurrency.lockutils [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "a2d6db04-9293-4755-ad8c-2e46467f92cb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.619399] env[63021]: DEBUG oslo_concurrency.lockutils [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "a2d6db04-9293-4755-ad8c-2e46467f92cb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.619572] env[63021]: DEBUG oslo_concurrency.lockutils [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "a2d6db04-9293-4755-ad8c-2e46467f92cb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.622536] env[63021]: INFO nova.compute.manager [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Terminating instance [ 1076.624199] env[63021]: DEBUG nova.compute.manager [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1076.624402] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1076.625245] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d7fec97-d3a5-4853-adda-7ce3963c20ec {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.632251] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1076.632484] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-21785bb0-cefa-45fc-a2a5-0dc1443854a7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.639825] env[63021]: DEBUG oslo_vmware.api [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1076.639825] env[63021]: value = "task-1294293" [ 1076.639825] env[63021]: _type = "Task" [ 1076.639825] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.647845] env[63021]: DEBUG oslo_vmware.api [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294293, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.907477] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-e571f476-5708-4f52-9bfd-679d05b0fca6 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Volume attach. Driver type: vmdk {{(pid=63021) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1076.907739] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-e571f476-5708-4f52-9bfd-679d05b0fca6 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277598', 'volume_id': 'cd524289-5d08-4b7f-a083-259e501f1fc1', 'name': 'volume-cd524289-5d08-4b7f-a083-259e501f1fc1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd03c47da-7e73-4474-ad9b-07d27629dcae', 'attached_at': '', 'detached_at': '', 'volume_id': 'cd524289-5d08-4b7f-a083-259e501f1fc1', 'serial': 'cd524289-5d08-4b7f-a083-259e501f1fc1'} {{(pid=63021) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1076.908727] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89170659-4a74-4cfb-9517-689c79a5f87a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.925996] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92e65f5e-f14d-4b2c-a995-65cadd754a61 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.952011] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-e571f476-5708-4f52-9bfd-679d05b0fca6 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] volume-cd524289-5d08-4b7f-a083-259e501f1fc1/volume-cd524289-5d08-4b7f-a083-259e501f1fc1.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1076.955984] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-de1d93b5-d586-460b-838b-1fbd5c3cbe4c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.975393] env[63021]: DEBUG oslo_vmware.api [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52d17cf6-e18d-5191-139d-a4fa567dad6e, 'name': SearchDatastore_Task, 'duration_secs': 0.01469} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.976576] env[63021]: DEBUG oslo_concurrency.lockutils [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1076.976891] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 4041dfbf-fa94-4ff8-9176-e7088cdf241f/4041dfbf-fa94-4ff8-9176-e7088cdf241f.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1076.977191] env[63021]: DEBUG oslo_vmware.api [None req-e571f476-5708-4f52-9bfd-679d05b0fca6 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1076.977191] env[63021]: value = "task-1294294" [ 1076.977191] env[63021]: _type = "Task" [ 1076.977191] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.977390] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-56911821-98de-467a-8b1d-2025b5411f1d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.988411] env[63021]: DEBUG oslo_vmware.api [None req-e571f476-5708-4f52-9bfd-679d05b0fca6 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294294, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.989560] env[63021]: DEBUG oslo_vmware.api [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1076.989560] env[63021]: value = "task-1294295" [ 1076.989560] env[63021]: _type = "Task" [ 1076.989560] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.997467] env[63021]: DEBUG oslo_vmware.api [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294295, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.150645] env[63021]: DEBUG oslo_vmware.api [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294293, 'name': PowerOffVM_Task, 'duration_secs': 0.196345} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.150966] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1077.151163] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1077.151441] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3c518ea8-d1e2-4c53-b7aa-c0fd2cee461a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.218340] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1077.218668] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1077.218868] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Deleting the datastore file [datastore1] a2d6db04-9293-4755-ad8c-2e46467f92cb {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1077.219258] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2f562361-c3c4-4049-8ded-f6a37c9c1420 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.226071] env[63021]: DEBUG oslo_vmware.api [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1077.226071] env[63021]: value = "task-1294297" [ 1077.226071] env[63021]: _type = "Task" [ 1077.226071] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.235348] env[63021]: DEBUG oslo_vmware.api [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294297, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.410403] env[63021]: DEBUG oslo_concurrency.lockutils [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "e940f625-f35e-459d-b8be-fa5916ce6efb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.410812] env[63021]: DEBUG oslo_concurrency.lockutils [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "e940f625-f35e-459d-b8be-fa5916ce6efb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.410907] env[63021]: DEBUG oslo_concurrency.lockutils [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "e940f625-f35e-459d-b8be-fa5916ce6efb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.489779] env[63021]: DEBUG oslo_vmware.api [None req-e571f476-5708-4f52-9bfd-679d05b0fca6 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294294, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.498034] env[63021]: DEBUG oslo_vmware.api [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294295, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.496972} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.498388] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore2] 4041dfbf-fa94-4ff8-9176-e7088cdf241f/4041dfbf-fa94-4ff8-9176-e7088cdf241f.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1077.498632] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1077.498894] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ebeec353-fb15-49f6-9a26-1fdc4aa8bee7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.504623] env[63021]: DEBUG oslo_vmware.api [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1077.504623] env[63021]: value = "task-1294298" [ 1077.504623] env[63021]: _type = "Task" [ 1077.504623] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.512104] env[63021]: DEBUG oslo_vmware.api [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294298, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.610183] env[63021]: DEBUG oslo_concurrency.lockutils [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "67f736d9-82ec-448e-8b72-84fdd704aa8d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.610462] env[63021]: DEBUG oslo_concurrency.lockutils [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "67f736d9-82ec-448e-8b72-84fdd704aa8d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.610698] env[63021]: DEBUG oslo_concurrency.lockutils [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "67f736d9-82ec-448e-8b72-84fdd704aa8d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.610879] env[63021]: DEBUG oslo_concurrency.lockutils [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "67f736d9-82ec-448e-8b72-84fdd704aa8d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.611278] env[63021]: DEBUG oslo_concurrency.lockutils [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "67f736d9-82ec-448e-8b72-84fdd704aa8d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.613255] env[63021]: INFO nova.compute.manager [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Terminating instance [ 1077.614969] env[63021]: DEBUG nova.compute.manager [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1077.615284] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1077.615990] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c009902b-00dd-4ae4-b8a6-9e63b3f8920a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.623729] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1077.623970] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dc41ae04-7ab0-4b68-a1c0-d34d035b6299 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.630318] env[63021]: DEBUG oslo_vmware.api [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 1077.630318] env[63021]: value = "task-1294299" [ 1077.630318] env[63021]: _type = "Task" [ 1077.630318] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.638188] env[63021]: DEBUG oslo_vmware.api [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294299, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.735774] env[63021]: DEBUG oslo_vmware.api [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294297, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.305565} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.736097] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1077.736241] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1077.736423] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1077.736624] env[63021]: INFO nova.compute.manager [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1077.736843] env[63021]: DEBUG oslo.service.loopingcall [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1077.737053] env[63021]: DEBUG nova.compute.manager [-] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1077.737214] env[63021]: DEBUG nova.network.neutron [-] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1077.996025] env[63021]: DEBUG oslo_vmware.api [None req-e571f476-5708-4f52-9bfd-679d05b0fca6 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294294, 'name': ReconfigVM_Task, 'duration_secs': 0.620175} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.996025] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-e571f476-5708-4f52-9bfd-679d05b0fca6 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Reconfigured VM instance instance-00000062 to attach disk [datastore2] volume-cd524289-5d08-4b7f-a083-259e501f1fc1/volume-cd524289-5d08-4b7f-a083-259e501f1fc1.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1077.998097] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-389aa8cd-c91b-44d6-b564-1cbe9168b5ec {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.022985] env[63021]: DEBUG oslo_vmware.api [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294298, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066393} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.024289] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1078.024912] env[63021]: DEBUG oslo_vmware.api [None req-e571f476-5708-4f52-9bfd-679d05b0fca6 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1078.024912] env[63021]: value = "task-1294301" [ 1078.024912] env[63021]: _type = "Task" [ 1078.024912] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.025966] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94339da2-9e57-4e14-9704-e8a1470b2a11 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.037762] env[63021]: DEBUG oslo_vmware.api [None req-e571f476-5708-4f52-9bfd-679d05b0fca6 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294301, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.056954] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] 4041dfbf-fa94-4ff8-9176-e7088cdf241f/4041dfbf-fa94-4ff8-9176-e7088cdf241f.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1078.059073] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cc9cb126-63b4-4dca-870a-4a538f828c84 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.075112] env[63021]: DEBUG nova.network.neutron [-] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.079553] env[63021]: DEBUG nova.compute.manager [req-1bebfc4a-aa03-4b00-9ce0-a0f7737762b0 req-824659e2-33df-43c9-bf50-e76dc989919e service nova] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Received event network-vif-deleted-ef03a045-2fd4-4794-964b-ec6a72e8f7d2 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1078.079854] env[63021]: INFO nova.compute.manager [req-1bebfc4a-aa03-4b00-9ce0-a0f7737762b0 req-824659e2-33df-43c9-bf50-e76dc989919e service nova] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Neutron deleted interface ef03a045-2fd4-4794-964b-ec6a72e8f7d2; detaching it from the instance and deleting it from the info cache [ 1078.080105] env[63021]: DEBUG nova.network.neutron [req-1bebfc4a-aa03-4b00-9ce0-a0f7737762b0 req-824659e2-33df-43c9-bf50-e76dc989919e service nova] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.083799] env[63021]: DEBUG oslo_vmware.api [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1078.083799] env[63021]: value = "task-1294302" [ 1078.083799] env[63021]: _type = "Task" [ 1078.083799] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.097722] env[63021]: DEBUG oslo_vmware.api [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294302, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.142039] env[63021]: DEBUG oslo_vmware.api [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294299, 'name': PowerOffVM_Task, 'duration_secs': 0.164957} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.142319] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1078.142553] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1078.142869] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5bb58bf3-b8f3-48a6-921a-7ed63a54c85d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.226458] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1078.226687] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1078.226882] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Deleting the datastore file [datastore1] 67f736d9-82ec-448e-8b72-84fdd704aa8d {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1078.227199] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2db40981-6f8b-4767-943d-bde45baad5e3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.234731] env[63021]: DEBUG oslo_vmware.api [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 1078.234731] env[63021]: value = "task-1294304" [ 1078.234731] env[63021]: _type = "Task" [ 1078.234731] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.242879] env[63021]: DEBUG oslo_vmware.api [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294304, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.446244] env[63021]: DEBUG oslo_concurrency.lockutils [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "refresh_cache-e940f625-f35e-459d-b8be-fa5916ce6efb" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1078.446437] env[63021]: DEBUG oslo_concurrency.lockutils [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquired lock "refresh_cache-e940f625-f35e-459d-b8be-fa5916ce6efb" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.446668] env[63021]: DEBUG nova.network.neutron [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1078.539715] env[63021]: DEBUG oslo_vmware.api [None req-e571f476-5708-4f52-9bfd-679d05b0fca6 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294301, 'name': ReconfigVM_Task, 'duration_secs': 0.148481} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.540060] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-e571f476-5708-4f52-9bfd-679d05b0fca6 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277598', 'volume_id': 'cd524289-5d08-4b7f-a083-259e501f1fc1', 'name': 'volume-cd524289-5d08-4b7f-a083-259e501f1fc1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd03c47da-7e73-4474-ad9b-07d27629dcae', 'attached_at': '', 'detached_at': '', 'volume_id': 'cd524289-5d08-4b7f-a083-259e501f1fc1', 'serial': 'cd524289-5d08-4b7f-a083-259e501f1fc1'} {{(pid=63021) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1078.583705] env[63021]: INFO nova.compute.manager [-] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Took 0.85 seconds to deallocate network for instance. [ 1078.586099] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-292ccd22-3e7c-41ce-ad69-7f7b01dbee44 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.600371] env[63021]: DEBUG oslo_vmware.api [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294302, 'name': ReconfigVM_Task, 'duration_secs': 0.278705} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.601715] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Reconfigured VM instance instance-00000069 to attach disk [datastore2] 4041dfbf-fa94-4ff8-9176-e7088cdf241f/4041dfbf-fa94-4ff8-9176-e7088cdf241f.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1078.602463] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f75842b4-22e3-46f5-ae9c-eefe2cf895a2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.607284] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e4af79e-3eae-4ed1-8386-643601613dc4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.629669] env[63021]: DEBUG oslo_vmware.api [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1078.629669] env[63021]: value = "task-1294305" [ 1078.629669] env[63021]: _type = "Task" [ 1078.629669] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.645874] env[63021]: DEBUG nova.compute.manager [req-1bebfc4a-aa03-4b00-9ce0-a0f7737762b0 req-824659e2-33df-43c9-bf50-e76dc989919e service nova] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Detach interface failed, port_id=ef03a045-2fd4-4794-964b-ec6a72e8f7d2, reason: Instance a2d6db04-9293-4755-ad8c-2e46467f92cb could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1078.651368] env[63021]: DEBUG oslo_vmware.api [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294305, 'name': Rename_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.744498] env[63021]: DEBUG oslo_vmware.api [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294304, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.209786} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.744780] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1078.744989] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1078.745250] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1078.745469] env[63021]: INFO nova.compute.manager [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1078.745717] env[63021]: DEBUG oslo.service.loopingcall [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1078.745937] env[63021]: DEBUG nova.compute.manager [-] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1078.746041] env[63021]: DEBUG nova.network.neutron [-] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1079.099190] env[63021]: DEBUG oslo_concurrency.lockutils [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.099529] env[63021]: DEBUG oslo_concurrency.lockutils [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.099701] env[63021]: DEBUG nova.objects.instance [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lazy-loading 'resources' on Instance uuid a2d6db04-9293-4755-ad8c-2e46467f92cb {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1079.146933] env[63021]: DEBUG oslo_vmware.api [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294305, 'name': Rename_Task, 'duration_secs': 0.208028} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.147172] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1079.147434] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e4d926e4-a23a-4408-b54f-c487c849161d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.155338] env[63021]: DEBUG oslo_vmware.api [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1079.155338] env[63021]: value = "task-1294306" [ 1079.155338] env[63021]: _type = "Task" [ 1079.155338] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.164271] env[63021]: DEBUG oslo_vmware.api [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294306, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.182529] env[63021]: DEBUG nova.network.neutron [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Updating instance_info_cache with network_info: [{"id": "d2876adc-ed3c-4540-b06d-fd179ecd2445", "address": "fa:16:3e:a2:13:81", "network": {"id": "81917a55-4e41-4898-9f4e-4fced52b5a86", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1793837736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e238d5c1668540d2a9fffd3fd832d9b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2876adc-ed", "ovs_interfaceid": "d2876adc-ed3c-4540-b06d-fd179ecd2445", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1079.669023] env[63021]: DEBUG oslo_vmware.api [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294306, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.682039] env[63021]: DEBUG nova.objects.instance [None req-e571f476-5708-4f52-9bfd-679d05b0fca6 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lazy-loading 'flavor' on Instance uuid d03c47da-7e73-4474-ad9b-07d27629dcae {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1079.685244] env[63021]: DEBUG oslo_concurrency.lockutils [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Releasing lock "refresh_cache-e940f625-f35e-459d-b8be-fa5916ce6efb" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1079.764800] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1b74f20-5e41-4600-8619-615f17597855 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.772377] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e74cea6-6a75-43ee-b256-7e19c7cf4c09 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.802614] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-286e8823-e788-4913-af48-36d74b693cd2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.809655] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8019a0e-8aca-4610-b04c-358c39a30276 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.824231] env[63021]: DEBUG nova.compute.provider_tree [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1079.894799] env[63021]: DEBUG nova.network.neutron [-] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1080.109678] env[63021]: DEBUG nova.compute.manager [req-c88bff42-d7ec-479b-858d-330ccd81f958 req-125ad79a-86fe-4a93-862b-735090665e43 service nova] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Received event network-vif-deleted-db333f88-b16f-439c-99fe-ca5884a09dc1 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1080.166520] env[63021]: DEBUG oslo_vmware.api [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294306, 'name': PowerOnVM_Task, 'duration_secs': 0.536131} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.166914] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1080.166914] env[63021]: INFO nova.compute.manager [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Took 7.60 seconds to spawn the instance on the hypervisor. [ 1080.167123] env[63021]: DEBUG nova.compute.manager [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1080.167942] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d78fafc4-e2c0-4ebc-abfe-1b7892f246c9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.189133] env[63021]: DEBUG oslo_concurrency.lockutils [None req-e571f476-5708-4f52-9bfd-679d05b0fca6 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "d03c47da-7e73-4474-ad9b-07d27629dcae" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.882s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.198301] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f365d8b-2d16-46bd-af71-e14f2b9a311c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Volume attach. Driver type: vmdk {{(pid=63021) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1080.198677] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f365d8b-2d16-46bd-af71-e14f2b9a311c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277600', 'volume_id': 'df640a1b-6005-45df-9ce3-d8634fcb4380', 'name': 'volume-df640a1b-6005-45df-9ce3-d8634fcb4380', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8d9387b8-6a84-4ada-8ac7-4ffd922b675a', 'attached_at': '', 'detached_at': '', 'volume_id': 'df640a1b-6005-45df-9ce3-d8634fcb4380', 'serial': 'df640a1b-6005-45df-9ce3-d8634fcb4380'} {{(pid=63021) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1080.200567] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6de30d5-ca1f-4494-8183-ce01964ed93a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.221990] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12d7e25d-106f-4733-8621-aa815140a153 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.225009] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e534c88-5b2c-4562-95f1-70f227168bd2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.260109] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffc0f0a7-ff02-4967-984c-edd848e4c14d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.270514] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f365d8b-2d16-46bd-af71-e14f2b9a311c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] volume-df640a1b-6005-45df-9ce3-d8634fcb4380/volume-df640a1b-6005-45df-9ce3-d8634fcb4380.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1080.270647] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ba41ccb4-71ad-4165-886d-4b7353933896 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.290015] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Updating instance 'e940f625-f35e-459d-b8be-fa5916ce6efb' progress to 83 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1080.297448] env[63021]: DEBUG oslo_vmware.api [None req-1f365d8b-2d16-46bd-af71-e14f2b9a311c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 1080.297448] env[63021]: value = "task-1294307" [ 1080.297448] env[63021]: _type = "Task" [ 1080.297448] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.304811] env[63021]: DEBUG oslo_vmware.api [None req-1f365d8b-2d16-46bd-af71-e14f2b9a311c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294307, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.329296] env[63021]: DEBUG nova.scheduler.client.report [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1080.397768] env[63021]: INFO nova.compute.manager [-] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Took 1.65 seconds to deallocate network for instance. [ 1080.689103] env[63021]: INFO nova.compute.manager [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Took 12.82 seconds to build instance. [ 1080.797944] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1080.798509] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f34df292-c8cc-49f1-a9bb-e418ef49606a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.810425] env[63021]: DEBUG oslo_vmware.api [None req-1f365d8b-2d16-46bd-af71-e14f2b9a311c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294307, 'name': ReconfigVM_Task, 'duration_secs': 0.428994} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.811687] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f365d8b-2d16-46bd-af71-e14f2b9a311c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Reconfigured VM instance instance-00000061 to attach disk [datastore2] volume-df640a1b-6005-45df-9ce3-d8634fcb4380/volume-df640a1b-6005-45df-9ce3-d8634fcb4380.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1080.816515] env[63021]: DEBUG oslo_vmware.api [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1080.816515] env[63021]: value = "task-1294308" [ 1080.816515] env[63021]: _type = "Task" [ 1080.816515] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.816737] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dd290ecc-23d3-4e36-91ae-c1c28d891224 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.837400] env[63021]: DEBUG oslo_concurrency.lockutils [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.738s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.840564] env[63021]: DEBUG oslo_vmware.api [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294308, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.843548] env[63021]: DEBUG oslo_vmware.api [None req-1f365d8b-2d16-46bd-af71-e14f2b9a311c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 1080.843548] env[63021]: value = "task-1294309" [ 1080.843548] env[63021]: _type = "Task" [ 1080.843548] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.851261] env[63021]: DEBUG oslo_vmware.api [None req-1f365d8b-2d16-46bd-af71-e14f2b9a311c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294309, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.860819] env[63021]: INFO nova.scheduler.client.report [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Deleted allocations for instance a2d6db04-9293-4755-ad8c-2e46467f92cb [ 1080.905503] env[63021]: DEBUG oslo_concurrency.lockutils [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.905793] env[63021]: DEBUG oslo_concurrency.lockutils [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.906159] env[63021]: DEBUG nova.objects.instance [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lazy-loading 'resources' on Instance uuid 67f736d9-82ec-448e-8b72-84fdd704aa8d {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1081.194054] env[63021]: DEBUG oslo_concurrency.lockutils [None req-92179f57-4496-4334-ac15-308701ac99e8 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "4041dfbf-fa94-4ff8-9176-e7088cdf241f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.335s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.337487] env[63021]: DEBUG oslo_vmware.api [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294308, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.351552] env[63021]: DEBUG oslo_vmware.api [None req-1f365d8b-2d16-46bd-af71-e14f2b9a311c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294309, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.369960] env[63021]: DEBUG oslo_concurrency.lockutils [None req-77721ae1-923d-424c-95a2-b80c7c627b80 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "a2d6db04-9293-4755-ad8c-2e46467f92cb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.751s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.549494] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-292d2c11-9620-457c-abd2-624a0b33804d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.556698] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0788ddfc-e90d-49d5-b5cd-2e0424fed181 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.587857] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ac35078-a6b1-439b-823a-c8b05effb83e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.594816] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08e5a95a-2db8-47e8-84d0-42d761eb927d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.607604] env[63021]: DEBUG nova.compute.provider_tree [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1081.838073] env[63021]: DEBUG oslo_vmware.api [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294308, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.851141] env[63021]: DEBUG oslo_vmware.api [None req-1f365d8b-2d16-46bd-af71-e14f2b9a311c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294309, 'name': ReconfigVM_Task, 'duration_secs': 0.981635} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.851482] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f365d8b-2d16-46bd-af71-e14f2b9a311c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277600', 'volume_id': 'df640a1b-6005-45df-9ce3-d8634fcb4380', 'name': 'volume-df640a1b-6005-45df-9ce3-d8634fcb4380', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8d9387b8-6a84-4ada-8ac7-4ffd922b675a', 'attached_at': '', 'detached_at': '', 'volume_id': 'df640a1b-6005-45df-9ce3-d8634fcb4380', 'serial': 'df640a1b-6005-45df-9ce3-d8634fcb4380'} {{(pid=63021) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1081.920603] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "02c217fb-174a-47f8-a9b7-d71497797590" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.920842] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "02c217fb-174a-47f8-a9b7-d71497797590" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.111357] env[63021]: DEBUG nova.scheduler.client.report [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1082.138848] env[63021]: DEBUG nova.compute.manager [req-2396a9a0-4656-4fe1-ae09-0f9c81be62b7 req-c1718994-ed5a-4117-8dd3-6b65d1455550 service nova] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Received event network-changed-b9d1b591-e9bc-4e70-9c76-1a0673bca1d5 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1082.139086] env[63021]: DEBUG nova.compute.manager [req-2396a9a0-4656-4fe1-ae09-0f9c81be62b7 req-c1718994-ed5a-4117-8dd3-6b65d1455550 service nova] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Refreshing instance network info cache due to event network-changed-b9d1b591-e9bc-4e70-9c76-1a0673bca1d5. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1082.139284] env[63021]: DEBUG oslo_concurrency.lockutils [req-2396a9a0-4656-4fe1-ae09-0f9c81be62b7 req-c1718994-ed5a-4117-8dd3-6b65d1455550 service nova] Acquiring lock "refresh_cache-4041dfbf-fa94-4ff8-9176-e7088cdf241f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1082.139433] env[63021]: DEBUG oslo_concurrency.lockutils [req-2396a9a0-4656-4fe1-ae09-0f9c81be62b7 req-c1718994-ed5a-4117-8dd3-6b65d1455550 service nova] Acquired lock "refresh_cache-4041dfbf-fa94-4ff8-9176-e7088cdf241f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.139596] env[63021]: DEBUG nova.network.neutron [req-2396a9a0-4656-4fe1-ae09-0f9c81be62b7 req-c1718994-ed5a-4117-8dd3-6b65d1455550 service nova] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Refreshing network info cache for port b9d1b591-e9bc-4e70-9c76-1a0673bca1d5 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1082.194246] env[63021]: DEBUG oslo_concurrency.lockutils [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "e2e98322-fdda-45e0-b3ba-58349ab940be" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.194575] env[63021]: DEBUG oslo_concurrency.lockutils [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "e2e98322-fdda-45e0-b3ba-58349ab940be" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.194715] env[63021]: DEBUG oslo_concurrency.lockutils [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "e2e98322-fdda-45e0-b3ba-58349ab940be-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.194904] env[63021]: DEBUG oslo_concurrency.lockutils [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "e2e98322-fdda-45e0-b3ba-58349ab940be-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.195095] env[63021]: DEBUG oslo_concurrency.lockutils [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "e2e98322-fdda-45e0-b3ba-58349ab940be-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.197265] env[63021]: INFO nova.compute.manager [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Terminating instance [ 1082.199075] env[63021]: DEBUG nova.compute.manager [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1082.199276] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1082.200138] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c4817ac-9dca-40dd-822c-e0b52586cc73 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.207927] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1082.208181] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e607c9bb-c7a8-4085-a56f-52456da9b132 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.214221] env[63021]: DEBUG oslo_vmware.api [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1082.214221] env[63021]: value = "task-1294310" [ 1082.214221] env[63021]: _type = "Task" [ 1082.214221] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.224050] env[63021]: DEBUG oslo_vmware.api [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294310, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.338164] env[63021]: DEBUG oslo_vmware.api [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294308, 'name': PowerOnVM_Task, 'duration_secs': 1.326293} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.338550] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1082.338775] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-62200fd0-62f6-481b-84d7-aefe85a96ca5 tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Updating instance 'e940f625-f35e-459d-b8be-fa5916ce6efb' progress to 100 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1082.422992] env[63021]: DEBUG nova.compute.manager [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1082.616879] env[63021]: DEBUG oslo_concurrency.lockutils [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.711s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.642046] env[63021]: INFO nova.scheduler.client.report [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Deleted allocations for instance 67f736d9-82ec-448e-8b72-84fdd704aa8d [ 1082.724965] env[63021]: DEBUG oslo_vmware.api [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294310, 'name': PowerOffVM_Task, 'duration_secs': 0.20984} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.725306] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1082.725510] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1082.725775] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b5dfb0bf-7920-4ddd-90a3-20e01914d64a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.891805] env[63021]: DEBUG nova.network.neutron [req-2396a9a0-4656-4fe1-ae09-0f9c81be62b7 req-c1718994-ed5a-4117-8dd3-6b65d1455550 service nova] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Updated VIF entry in instance network info cache for port b9d1b591-e9bc-4e70-9c76-1a0673bca1d5. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1082.892261] env[63021]: DEBUG nova.network.neutron [req-2396a9a0-4656-4fe1-ae09-0f9c81be62b7 req-c1718994-ed5a-4117-8dd3-6b65d1455550 service nova] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Updating instance_info_cache with network_info: [{"id": "b9d1b591-e9bc-4e70-9c76-1a0673bca1d5", "address": "fa:16:3e:1b:b0:e6", "network": {"id": "9502e364-60e1-47ef-a1f7-f3e3ff45cf69", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-164717784-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5af87713ae6b477ca8a584aed287889d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9d1b591-e9", "ovs_interfaceid": "b9d1b591-e9bc-4e70-9c76-1a0673bca1d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.893930] env[63021]: DEBUG nova.objects.instance [None req-1f365d8b-2d16-46bd-af71-e14f2b9a311c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lazy-loading 'flavor' on Instance uuid 8d9387b8-6a84-4ada-8ac7-4ffd922b675a {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1082.945070] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.945429] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.947014] env[63021]: INFO nova.compute.claims [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1083.151453] env[63021]: DEBUG oslo_concurrency.lockutils [None req-943a6489-1517-4f77-9bfa-e5b04c8af9ce tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "67f736d9-82ec-448e-8b72-84fdd704aa8d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.541s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.330877] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1083.331201] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1083.331320] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Deleting the datastore file [datastore1] e2e98322-fdda-45e0-b3ba-58349ab940be {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1083.331603] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a4e93e38-19d1-464f-8b87-51a62c12507f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.338540] env[63021]: DEBUG oslo_vmware.api [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1083.338540] env[63021]: value = "task-1294312" [ 1083.338540] env[63021]: _type = "Task" [ 1083.338540] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.346623] env[63021]: DEBUG oslo_vmware.api [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294312, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.396044] env[63021]: DEBUG oslo_concurrency.lockutils [req-2396a9a0-4656-4fe1-ae09-0f9c81be62b7 req-c1718994-ed5a-4117-8dd3-6b65d1455550 service nova] Releasing lock "refresh_cache-4041dfbf-fa94-4ff8-9176-e7088cdf241f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.399023] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1f365d8b-2d16-46bd-af71-e14f2b9a311c tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "8d9387b8-6a84-4ada-8ac7-4ffd922b675a" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.804s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.848755] env[63021]: DEBUG oslo_vmware.api [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294312, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141428} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.849220] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1083.849617] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1083.849925] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1083.850293] env[63021]: INFO nova.compute.manager [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Took 1.65 seconds to destroy the instance on the hypervisor. [ 1083.850711] env[63021]: DEBUG oslo.service.loopingcall [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1083.852066] env[63021]: DEBUG nova.compute.manager [-] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1083.852308] env[63021]: DEBUG nova.network.neutron [-] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1084.136866] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4700df6b-a156-4d17-be2f-ca586a3341c0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.143813] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86ec77af-355c-4033-a8e4-94300ea387fd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.146934] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f1db2fea-d792-457a-8b82-430d36a2f058 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquiring lock "8d9387b8-6a84-4ada-8ac7-4ffd922b675a" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.147185] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f1db2fea-d792-457a-8b82-430d36a2f058 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "8d9387b8-6a84-4ada-8ac7-4ffd922b675a" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.177093] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f73e1ad6-1fa7-4cf1-ad6a-ad6c41829a08 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.180529] env[63021]: DEBUG nova.compute.manager [req-1db4d089-7b88-4d52-8080-93f39aadf6fc req-06fa9fda-5838-4f4a-adde-cf1c3064031e service nova] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Received event network-vif-deleted-b9cfa391-5cd3-4e10-9ee1-547f1ee41769 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1084.180677] env[63021]: INFO nova.compute.manager [req-1db4d089-7b88-4d52-8080-93f39aadf6fc req-06fa9fda-5838-4f4a-adde-cf1c3064031e service nova] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Neutron deleted interface b9cfa391-5cd3-4e10-9ee1-547f1ee41769; detaching it from the instance and deleting it from the info cache [ 1084.180853] env[63021]: DEBUG nova.network.neutron [req-1db4d089-7b88-4d52-8080-93f39aadf6fc req-06fa9fda-5838-4f4a-adde-cf1c3064031e service nova] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.187912] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caacabd9-90c6-46de-b94b-5883be7a2a6d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.203704] env[63021]: DEBUG nova.compute.provider_tree [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1084.261141] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "e940f625-f35e-459d-b8be-fa5916ce6efb" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.261529] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "e940f625-f35e-459d-b8be-fa5916ce6efb" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.261765] env[63021]: DEBUG nova.compute.manager [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Going to confirm migration 4 {{(pid=63021) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1084.625127] env[63021]: DEBUG nova.network.neutron [-] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.650573] env[63021]: INFO nova.compute.manager [None req-f1db2fea-d792-457a-8b82-430d36a2f058 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Detaching volume 29440f30-19f1-42a5-9503-a29373b566ff [ 1084.682960] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e3da04fd-3e00-4950-af32-75e7765ec2ad {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.685970] env[63021]: INFO nova.virt.block_device [None req-f1db2fea-d792-457a-8b82-430d36a2f058 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Attempting to driver detach volume 29440f30-19f1-42a5-9503-a29373b566ff from mountpoint /dev/sdb [ 1084.686248] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1db2fea-d792-457a-8b82-430d36a2f058 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Volume detach. Driver type: vmdk {{(pid=63021) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1084.686441] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1db2fea-d792-457a-8b82-430d36a2f058 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277595', 'volume_id': '29440f30-19f1-42a5-9503-a29373b566ff', 'name': 'volume-29440f30-19f1-42a5-9503-a29373b566ff', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8d9387b8-6a84-4ada-8ac7-4ffd922b675a', 'attached_at': '', 'detached_at': '', 'volume_id': '29440f30-19f1-42a5-9503-a29373b566ff', 'serial': '29440f30-19f1-42a5-9503-a29373b566ff'} {{(pid=63021) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1084.687323] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7149711-af39-4db9-a09f-9df46637d110 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.711619] env[63021]: DEBUG nova.scheduler.client.report [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1084.716692] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2982b2c3-493f-4510-98fc-d405b0bebeea {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.721983] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee49dde2-47f1-4129-9520-c45a278efd44 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.751894] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80fb098f-971b-4b83-8982-9aafa3b857ab {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.754725] env[63021]: DEBUG nova.compute.manager [req-1db4d089-7b88-4d52-8080-93f39aadf6fc req-06fa9fda-5838-4f4a-adde-cf1c3064031e service nova] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Detach interface failed, port_id=b9cfa391-5cd3-4e10-9ee1-547f1ee41769, reason: Instance e2e98322-fdda-45e0-b3ba-58349ab940be could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1084.780501] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9a52120-530a-4356-9559-3fa51b5666c2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.795221] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1db2fea-d792-457a-8b82-430d36a2f058 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] The volume has not been displaced from its original location: [datastore1] volume-29440f30-19f1-42a5-9503-a29373b566ff/volume-29440f30-19f1-42a5-9503-a29373b566ff.vmdk. No consolidation needed. {{(pid=63021) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1084.800397] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1db2fea-d792-457a-8b82-430d36a2f058 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Reconfiguring VM instance instance-00000061 to detach disk 2001 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1084.800689] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c79fe6d-74ee-415e-9c6a-e3c8240ddfdf {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.813966] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "refresh_cache-e940f625-f35e-459d-b8be-fa5916ce6efb" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.814162] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquired lock "refresh_cache-e940f625-f35e-459d-b8be-fa5916ce6efb" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.814340] env[63021]: DEBUG nova.network.neutron [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1084.814526] env[63021]: DEBUG nova.objects.instance [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lazy-loading 'info_cache' on Instance uuid e940f625-f35e-459d-b8be-fa5916ce6efb {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1084.821537] env[63021]: DEBUG oslo_vmware.api [None req-f1db2fea-d792-457a-8b82-430d36a2f058 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 1084.821537] env[63021]: value = "task-1294314" [ 1084.821537] env[63021]: _type = "Task" [ 1084.821537] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.829392] env[63021]: DEBUG oslo_vmware.api [None req-f1db2fea-d792-457a-8b82-430d36a2f058 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294314, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.128064] env[63021]: INFO nova.compute.manager [-] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Took 1.28 seconds to deallocate network for instance. [ 1085.216197] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.271s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.216677] env[63021]: DEBUG nova.compute.manager [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1085.331509] env[63021]: DEBUG oslo_vmware.api [None req-f1db2fea-d792-457a-8b82-430d36a2f058 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294314, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.634921] env[63021]: DEBUG oslo_concurrency.lockutils [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.635319] env[63021]: DEBUG oslo_concurrency.lockutils [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.635491] env[63021]: DEBUG nova.objects.instance [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lazy-loading 'resources' on Instance uuid e2e98322-fdda-45e0-b3ba-58349ab940be {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1085.721491] env[63021]: DEBUG nova.compute.utils [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1085.722815] env[63021]: DEBUG nova.compute.manager [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1085.722983] env[63021]: DEBUG nova.network.neutron [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1085.766807] env[63021]: DEBUG nova.policy [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5431727705ff4c5483cd9a5b9cd104f6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '547d9a3099b94ff488ef6929ae591bac', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 1085.831639] env[63021]: DEBUG oslo_vmware.api [None req-f1db2fea-d792-457a-8b82-430d36a2f058 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294314, 'name': ReconfigVM_Task, 'duration_secs': 0.556223} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.833051] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1db2fea-d792-457a-8b82-430d36a2f058 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Reconfigured VM instance instance-00000061 to detach disk 2001 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1085.836561] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-80cee235-191b-46c0-90df-75fd27bbcf96 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.850722] env[63021]: DEBUG oslo_vmware.api [None req-f1db2fea-d792-457a-8b82-430d36a2f058 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 1085.850722] env[63021]: value = "task-1294315" [ 1085.850722] env[63021]: _type = "Task" [ 1085.850722] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.858430] env[63021]: DEBUG oslo_vmware.api [None req-f1db2fea-d792-457a-8b82-430d36a2f058 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294315, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.017974] env[63021]: DEBUG nova.network.neutron [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Successfully created port: 416fe99a-8969-4bb8-90d8-b9c9708917ca {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1086.020686] env[63021]: DEBUG nova.network.neutron [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Updating instance_info_cache with network_info: [{"id": "d2876adc-ed3c-4540-b06d-fd179ecd2445", "address": "fa:16:3e:a2:13:81", "network": {"id": "81917a55-4e41-4898-9f4e-4fced52b5a86", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1793837736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e238d5c1668540d2a9fffd3fd832d9b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2876adc-ed", "ovs_interfaceid": "d2876adc-ed3c-4540-b06d-fd179ecd2445", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1086.226535] env[63021]: DEBUG nova.compute.manager [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1086.278215] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4903915-3856-445d-bed2-2189ee4fcd92 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.287385] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad5a07a5-76cc-4f8e-8c99-cbf409aa1749 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.317021] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a01b923-6509-4eb4-b633-6b16672c6e4b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.324416] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b04ed98d-c9f8-41e0-91a5-65984fe7e1fb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.336792] env[63021]: DEBUG nova.compute.provider_tree [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1086.358489] env[63021]: DEBUG oslo_vmware.api [None req-f1db2fea-d792-457a-8b82-430d36a2f058 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294315, 'name': ReconfigVM_Task, 'duration_secs': 0.132935} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.358731] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1db2fea-d792-457a-8b82-430d36a2f058 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277595', 'volume_id': '29440f30-19f1-42a5-9503-a29373b566ff', 'name': 'volume-29440f30-19f1-42a5-9503-a29373b566ff', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8d9387b8-6a84-4ada-8ac7-4ffd922b675a', 'attached_at': '', 'detached_at': '', 'volume_id': '29440f30-19f1-42a5-9503-a29373b566ff', 'serial': '29440f30-19f1-42a5-9503-a29373b566ff'} {{(pid=63021) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1086.523394] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Releasing lock "refresh_cache-e940f625-f35e-459d-b8be-fa5916ce6efb" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.523706] env[63021]: DEBUG nova.objects.instance [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lazy-loading 'migration_context' on Instance uuid e940f625-f35e-459d-b8be-fa5916ce6efb {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1086.839876] env[63021]: DEBUG nova.scheduler.client.report [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1086.899356] env[63021]: DEBUG nova.objects.instance [None req-f1db2fea-d792-457a-8b82-430d36a2f058 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lazy-loading 'flavor' on Instance uuid 8d9387b8-6a84-4ada-8ac7-4ffd922b675a {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1087.026917] env[63021]: DEBUG nova.objects.base [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=63021) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1087.027940] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14cbaf8b-792c-4927-bbda-2183532dd8f3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.048594] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6de7e184-e4ed-4697-b690-1becd02f2437 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.054058] env[63021]: DEBUG oslo_vmware.api [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1087.054058] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]524c1b0d-892e-3e57-5c4e-8b3f4abdbe2b" [ 1087.054058] env[63021]: _type = "Task" [ 1087.054058] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.061539] env[63021]: DEBUG oslo_vmware.api [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]524c1b0d-892e-3e57-5c4e-8b3f4abdbe2b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.236632] env[63021]: DEBUG nova.compute.manager [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1087.263607] env[63021]: DEBUG nova.virt.hardware [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1087.263858] env[63021]: DEBUG nova.virt.hardware [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1087.264028] env[63021]: DEBUG nova.virt.hardware [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1087.264218] env[63021]: DEBUG nova.virt.hardware [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1087.264368] env[63021]: DEBUG nova.virt.hardware [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1087.264514] env[63021]: DEBUG nova.virt.hardware [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1087.264724] env[63021]: DEBUG nova.virt.hardware [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1087.264882] env[63021]: DEBUG nova.virt.hardware [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1087.265065] env[63021]: DEBUG nova.virt.hardware [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1087.265255] env[63021]: DEBUG nova.virt.hardware [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1087.265445] env[63021]: DEBUG nova.virt.hardware [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1087.266316] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c8cb1c3-91c7-48fe-8b64-0de8e2675751 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.274812] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3f708c0-2209-4332-ab6b-18a93a211686 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.344649] env[63021]: DEBUG oslo_concurrency.lockutils [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.709s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.367359] env[63021]: INFO nova.scheduler.client.report [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Deleted allocations for instance e2e98322-fdda-45e0-b3ba-58349ab940be [ 1087.441043] env[63021]: DEBUG nova.compute.manager [req-70ee648e-9de1-430a-b5f6-7da414774f40 req-1797f418-e6d4-4551-bb0b-9b2d155f3652 service nova] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Received event network-vif-plugged-416fe99a-8969-4bb8-90d8-b9c9708917ca {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1087.441279] env[63021]: DEBUG oslo_concurrency.lockutils [req-70ee648e-9de1-430a-b5f6-7da414774f40 req-1797f418-e6d4-4551-bb0b-9b2d155f3652 service nova] Acquiring lock "02c217fb-174a-47f8-a9b7-d71497797590-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.441494] env[63021]: DEBUG oslo_concurrency.lockutils [req-70ee648e-9de1-430a-b5f6-7da414774f40 req-1797f418-e6d4-4551-bb0b-9b2d155f3652 service nova] Lock "02c217fb-174a-47f8-a9b7-d71497797590-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.441660] env[63021]: DEBUG oslo_concurrency.lockutils [req-70ee648e-9de1-430a-b5f6-7da414774f40 req-1797f418-e6d4-4551-bb0b-9b2d155f3652 service nova] Lock "02c217fb-174a-47f8-a9b7-d71497797590-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.441857] env[63021]: DEBUG nova.compute.manager [req-70ee648e-9de1-430a-b5f6-7da414774f40 req-1797f418-e6d4-4551-bb0b-9b2d155f3652 service nova] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] No waiting events found dispatching network-vif-plugged-416fe99a-8969-4bb8-90d8-b9c9708917ca {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1087.441998] env[63021]: WARNING nova.compute.manager [req-70ee648e-9de1-430a-b5f6-7da414774f40 req-1797f418-e6d4-4551-bb0b-9b2d155f3652 service nova] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Received unexpected event network-vif-plugged-416fe99a-8969-4bb8-90d8-b9c9708917ca for instance with vm_state building and task_state spawning. [ 1087.532594] env[63021]: DEBUG nova.network.neutron [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Successfully updated port: 416fe99a-8969-4bb8-90d8-b9c9708917ca {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1087.564479] env[63021]: DEBUG oslo_vmware.api [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]524c1b0d-892e-3e57-5c4e-8b3f4abdbe2b, 'name': SearchDatastore_Task, 'duration_secs': 0.008229} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.564788] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.565031] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.874900] env[63021]: DEBUG oslo_concurrency.lockutils [None req-65fd3f08-a8c1-4eb0-85e6-21d9f4d14a0e tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "e2e98322-fdda-45e0-b3ba-58349ab940be" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.680s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.907111] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f1db2fea-d792-457a-8b82-430d36a2f058 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "8d9387b8-6a84-4ada-8ac7-4ffd922b675a" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.760s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.038967] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "refresh_cache-02c217fb-174a-47f8-a9b7-d71497797590" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1088.039156] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquired lock "refresh_cache-02c217fb-174a-47f8-a9b7-d71497797590" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1088.039337] env[63021]: DEBUG nova.network.neutron [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1088.206206] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7cdf996-7aa9-4226-b9ce-3c12f9d6a7f0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.214082] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b835dac-dcd3-48d1-9cb4-7baee6d2d83e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.244909] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ce07ba3-ab49-4ab1-91dd-0fb3935eb09d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.252611] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e22c2cf5-b839-40a9-9203-577b0fb652a3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.265675] env[63021]: DEBUG nova.compute.provider_tree [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1088.448897] env[63021]: DEBUG oslo_concurrency.lockutils [None req-eb061d70-dbcb-4a53-8d2a-c0b6d094c646 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquiring lock "8d9387b8-6a84-4ada-8ac7-4ffd922b675a" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.449204] env[63021]: DEBUG oslo_concurrency.lockutils [None req-eb061d70-dbcb-4a53-8d2a-c0b6d094c646 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "8d9387b8-6a84-4ada-8ac7-4ffd922b675a" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.564428] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Acquiring lock "d32d9644-de8b-464e-ae16-7fa775a622ee" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.564751] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Lock "d32d9644-de8b-464e-ae16-7fa775a622ee" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.565018] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Acquiring lock "d32d9644-de8b-464e-ae16-7fa775a622ee-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.565281] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Lock "d32d9644-de8b-464e-ae16-7fa775a622ee-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.565473] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Lock "d32d9644-de8b-464e-ae16-7fa775a622ee-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.567913] env[63021]: INFO nova.compute.manager [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Terminating instance [ 1088.569839] env[63021]: DEBUG nova.compute.manager [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1088.570073] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1088.571174] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac1a2633-fc6a-4ec3-ba76-2b8cd02da7e1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.579595] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1088.579849] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a5aef447-7e67-4ce2-aa89-6599a7d53204 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.583886] env[63021]: DEBUG nova.network.neutron [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1088.586718] env[63021]: DEBUG oslo_vmware.api [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Waiting for the task: (returnval){ [ 1088.586718] env[63021]: value = "task-1294316" [ 1088.586718] env[63021]: _type = "Task" [ 1088.586718] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.596629] env[63021]: DEBUG oslo_vmware.api [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Task: {'id': task-1294316, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.769462] env[63021]: DEBUG nova.scheduler.client.report [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1088.776021] env[63021]: DEBUG oslo_concurrency.lockutils [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "418a394b-dbb2-47a2-96a9-a1caa4c4bb51" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.776021] env[63021]: DEBUG oslo_concurrency.lockutils [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "418a394b-dbb2-47a2-96a9-a1caa4c4bb51" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.776547] env[63021]: DEBUG oslo_concurrency.lockutils [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "418a394b-dbb2-47a2-96a9-a1caa4c4bb51-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.776547] env[63021]: DEBUG oslo_concurrency.lockutils [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "418a394b-dbb2-47a2-96a9-a1caa4c4bb51-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.778630] env[63021]: DEBUG oslo_concurrency.lockutils [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "418a394b-dbb2-47a2-96a9-a1caa4c4bb51-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.779514] env[63021]: INFO nova.compute.manager [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Terminating instance [ 1088.781593] env[63021]: DEBUG nova.compute.manager [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1088.781755] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1088.782573] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d32964e-63a4-40fa-ba63-f8dfe5f92df5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.785979] env[63021]: DEBUG nova.network.neutron [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Updating instance_info_cache with network_info: [{"id": "416fe99a-8969-4bb8-90d8-b9c9708917ca", "address": "fa:16:3e:71:2a:8b", "network": {"id": "d27bb3cc-fb80-4b03-9288-ed17a0d05140", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-122910456-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "547d9a3099b94ff488ef6929ae591bac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1c797172-a569-458e-aeb0-3f21e589a740", "external-id": "nsx-vlan-transportzone-957", "segmentation_id": 957, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap416fe99a-89", "ovs_interfaceid": "416fe99a-8969-4bb8-90d8-b9c9708917ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1088.791979] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1088.792237] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b847aa6a-1603-4c63-a77f-57627934edcc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.797581] env[63021]: DEBUG oslo_vmware.api [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 1088.797581] env[63021]: value = "task-1294317" [ 1088.797581] env[63021]: _type = "Task" [ 1088.797581] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.806024] env[63021]: DEBUG oslo_vmware.api [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294317, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.952730] env[63021]: INFO nova.compute.manager [None req-eb061d70-dbcb-4a53-8d2a-c0b6d094c646 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Detaching volume df640a1b-6005-45df-9ce3-d8634fcb4380 [ 1088.990958] env[63021]: INFO nova.virt.block_device [None req-eb061d70-dbcb-4a53-8d2a-c0b6d094c646 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Attempting to driver detach volume df640a1b-6005-45df-9ce3-d8634fcb4380 from mountpoint /dev/sdc [ 1088.991325] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb061d70-dbcb-4a53-8d2a-c0b6d094c646 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Volume detach. Driver type: vmdk {{(pid=63021) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1088.991578] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb061d70-dbcb-4a53-8d2a-c0b6d094c646 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277600', 'volume_id': 'df640a1b-6005-45df-9ce3-d8634fcb4380', 'name': 'volume-df640a1b-6005-45df-9ce3-d8634fcb4380', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8d9387b8-6a84-4ada-8ac7-4ffd922b675a', 'attached_at': '', 'detached_at': '', 'volume_id': 'df640a1b-6005-45df-9ce3-d8634fcb4380', 'serial': 'df640a1b-6005-45df-9ce3-d8634fcb4380'} {{(pid=63021) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1088.992476] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a004f8f-a7b4-4d83-b7f9-ee6e6caa6aa2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.014136] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acbf82b8-d83d-4339-a7bf-c4e3298d4096 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.021010] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a92ee67e-8a6c-4ba9-ba70-d1b56f0dec76 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.040731] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8cc4710-50f5-48fa-837e-a61bbe8840dd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.055041] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb061d70-dbcb-4a53-8d2a-c0b6d094c646 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] The volume has not been displaced from its original location: [datastore2] volume-df640a1b-6005-45df-9ce3-d8634fcb4380/volume-df640a1b-6005-45df-9ce3-d8634fcb4380.vmdk. No consolidation needed. {{(pid=63021) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1089.060291] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb061d70-dbcb-4a53-8d2a-c0b6d094c646 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Reconfiguring VM instance instance-00000061 to detach disk 2002 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1089.060603] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6d47bdb3-2790-4345-850a-93e2ee3bfae7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.078395] env[63021]: DEBUG oslo_vmware.api [None req-eb061d70-dbcb-4a53-8d2a-c0b6d094c646 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 1089.078395] env[63021]: value = "task-1294318" [ 1089.078395] env[63021]: _type = "Task" [ 1089.078395] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.086395] env[63021]: DEBUG oslo_vmware.api [None req-eb061d70-dbcb-4a53-8d2a-c0b6d094c646 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294318, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.096254] env[63021]: DEBUG oslo_vmware.api [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Task: {'id': task-1294316, 'name': PowerOffVM_Task, 'duration_secs': 0.196922} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.096540] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1089.096717] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1089.096962] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-aefad4fb-2461-4150-aa6d-63ac8164768e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.165826] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1089.166107] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1089.166307] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Deleting the datastore file [datastore1] d32d9644-de8b-464e-ae16-7fa775a622ee {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1089.166593] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f4d633e8-686b-4505-adae-08ff9c2c1304 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.173354] env[63021]: DEBUG oslo_vmware.api [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Waiting for the task: (returnval){ [ 1089.173354] env[63021]: value = "task-1294320" [ 1089.173354] env[63021]: _type = "Task" [ 1089.173354] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.181040] env[63021]: DEBUG oslo_vmware.api [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Task: {'id': task-1294320, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.288889] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Releasing lock "refresh_cache-02c217fb-174a-47f8-a9b7-d71497797590" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1089.289342] env[63021]: DEBUG nova.compute.manager [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Instance network_info: |[{"id": "416fe99a-8969-4bb8-90d8-b9c9708917ca", "address": "fa:16:3e:71:2a:8b", "network": {"id": "d27bb3cc-fb80-4b03-9288-ed17a0d05140", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-122910456-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "547d9a3099b94ff488ef6929ae591bac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1c797172-a569-458e-aeb0-3f21e589a740", "external-id": "nsx-vlan-transportzone-957", "segmentation_id": 957, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap416fe99a-89", "ovs_interfaceid": "416fe99a-8969-4bb8-90d8-b9c9708917ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1089.289619] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:71:2a:8b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1c797172-a569-458e-aeb0-3f21e589a740', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '416fe99a-8969-4bb8-90d8-b9c9708917ca', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1089.297526] env[63021]: DEBUG oslo.service.loopingcall [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1089.298528] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1089.298892] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b7d15fc1-62e9-4701-a415-01f3dde44cf0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.328218] env[63021]: DEBUG oslo_vmware.api [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294317, 'name': PowerOffVM_Task, 'duration_secs': 0.224302} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.329597] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1089.329807] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1089.330093] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1089.330093] env[63021]: value = "task-1294321" [ 1089.330093] env[63021]: _type = "Task" [ 1089.330093] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.330271] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b135eeae-d939-43e2-85ca-35d51defc3dd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.340049] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294321, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.401438] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "ae8a512d-5393-432c-bca0-4dc86711891b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.401836] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "ae8a512d-5393-432c-bca0-4dc86711891b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.413152] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1089.413410] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1089.413629] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Deleting the datastore file [datastore1] 418a394b-dbb2-47a2-96a9-a1caa4c4bb51 {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1089.414543] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a2627653-73f8-4bb1-bbcb-b0d7755b6a30 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.421315] env[63021]: DEBUG oslo_vmware.api [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 1089.421315] env[63021]: value = "task-1294323" [ 1089.421315] env[63021]: _type = "Task" [ 1089.421315] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.430097] env[63021]: DEBUG oslo_vmware.api [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294323, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.512505] env[63021]: DEBUG nova.compute.manager [req-73144d8a-cb39-4707-990d-cf182ebd6cae req-ece7ccf9-7c0e-4870-8f84-85971e677c5b service nova] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Received event network-changed-416fe99a-8969-4bb8-90d8-b9c9708917ca {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1089.512664] env[63021]: DEBUG nova.compute.manager [req-73144d8a-cb39-4707-990d-cf182ebd6cae req-ece7ccf9-7c0e-4870-8f84-85971e677c5b service nova] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Refreshing instance network info cache due to event network-changed-416fe99a-8969-4bb8-90d8-b9c9708917ca. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1089.512869] env[63021]: DEBUG oslo_concurrency.lockutils [req-73144d8a-cb39-4707-990d-cf182ebd6cae req-ece7ccf9-7c0e-4870-8f84-85971e677c5b service nova] Acquiring lock "refresh_cache-02c217fb-174a-47f8-a9b7-d71497797590" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1089.513390] env[63021]: DEBUG oslo_concurrency.lockutils [req-73144d8a-cb39-4707-990d-cf182ebd6cae req-ece7ccf9-7c0e-4870-8f84-85971e677c5b service nova] Acquired lock "refresh_cache-02c217fb-174a-47f8-a9b7-d71497797590" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.513579] env[63021]: DEBUG nova.network.neutron [req-73144d8a-cb39-4707-990d-cf182ebd6cae req-ece7ccf9-7c0e-4870-8f84-85971e677c5b service nova] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Refreshing network info cache for port 416fe99a-8969-4bb8-90d8-b9c9708917ca {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1089.588643] env[63021]: DEBUG oslo_vmware.api [None req-eb061d70-dbcb-4a53-8d2a-c0b6d094c646 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294318, 'name': ReconfigVM_Task, 'duration_secs': 0.319279} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.588971] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb061d70-dbcb-4a53-8d2a-c0b6d094c646 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Reconfigured VM instance instance-00000061 to detach disk 2002 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1089.593745] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eb356639-3952-474d-9862-415430f5d5f5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.608450] env[63021]: DEBUG oslo_vmware.api [None req-eb061d70-dbcb-4a53-8d2a-c0b6d094c646 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 1089.608450] env[63021]: value = "task-1294324" [ 1089.608450] env[63021]: _type = "Task" [ 1089.608450] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.616381] env[63021]: DEBUG oslo_vmware.api [None req-eb061d70-dbcb-4a53-8d2a-c0b6d094c646 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294324, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.683753] env[63021]: DEBUG oslo_vmware.api [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Task: {'id': task-1294320, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.156213} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.684059] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1089.684261] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1089.684443] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1089.684622] env[63021]: INFO nova.compute.manager [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1089.684870] env[63021]: DEBUG oslo.service.loopingcall [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1089.685077] env[63021]: DEBUG nova.compute.manager [-] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1089.685175] env[63021]: DEBUG nova.network.neutron [-] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1089.780506] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.215s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.841650] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294321, 'name': CreateVM_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.904342] env[63021]: DEBUG nova.compute.manager [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1089.933027] env[63021]: DEBUG oslo_vmware.api [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294323, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.194531} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.933027] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1089.933027] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1089.933254] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1089.933362] env[63021]: INFO nova.compute.manager [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1089.933604] env[63021]: DEBUG oslo.service.loopingcall [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1089.933847] env[63021]: DEBUG nova.compute.manager [-] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1089.933908] env[63021]: DEBUG nova.network.neutron [-] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1090.119729] env[63021]: DEBUG oslo_vmware.api [None req-eb061d70-dbcb-4a53-8d2a-c0b6d094c646 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294324, 'name': ReconfigVM_Task, 'duration_secs': 0.145572} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.120070] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb061d70-dbcb-4a53-8d2a-c0b6d094c646 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277600', 'volume_id': 'df640a1b-6005-45df-9ce3-d8634fcb4380', 'name': 'volume-df640a1b-6005-45df-9ce3-d8634fcb4380', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8d9387b8-6a84-4ada-8ac7-4ffd922b675a', 'attached_at': '', 'detached_at': '', 'volume_id': 'df640a1b-6005-45df-9ce3-d8634fcb4380', 'serial': 'df640a1b-6005-45df-9ce3-d8634fcb4380'} {{(pid=63021) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1090.294274] env[63021]: DEBUG nova.compute.manager [req-f209ec01-eed0-40fe-8497-954f9842178f req-c9664a47-97e5-4dc4-b5a7-e4a3deb0eaf4 service nova] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Received event network-vif-deleted-905466dc-1047-469f-b282-768293b6d154 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1090.294483] env[63021]: INFO nova.compute.manager [req-f209ec01-eed0-40fe-8497-954f9842178f req-c9664a47-97e5-4dc4-b5a7-e4a3deb0eaf4 service nova] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Neutron deleted interface 905466dc-1047-469f-b282-768293b6d154; detaching it from the instance and deleting it from the info cache [ 1090.294690] env[63021]: DEBUG nova.network.neutron [req-f209ec01-eed0-40fe-8497-954f9842178f req-c9664a47-97e5-4dc4-b5a7-e4a3deb0eaf4 service nova] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.328622] env[63021]: DEBUG nova.network.neutron [req-73144d8a-cb39-4707-990d-cf182ebd6cae req-ece7ccf9-7c0e-4870-8f84-85971e677c5b service nova] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Updated VIF entry in instance network info cache for port 416fe99a-8969-4bb8-90d8-b9c9708917ca. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1090.328991] env[63021]: DEBUG nova.network.neutron [req-73144d8a-cb39-4707-990d-cf182ebd6cae req-ece7ccf9-7c0e-4870-8f84-85971e677c5b service nova] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Updating instance_info_cache with network_info: [{"id": "416fe99a-8969-4bb8-90d8-b9c9708917ca", "address": "fa:16:3e:71:2a:8b", "network": {"id": "d27bb3cc-fb80-4b03-9288-ed17a0d05140", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-122910456-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "547d9a3099b94ff488ef6929ae591bac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1c797172-a569-458e-aeb0-3f21e589a740", "external-id": "nsx-vlan-transportzone-957", "segmentation_id": 957, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap416fe99a-89", "ovs_interfaceid": "416fe99a-8969-4bb8-90d8-b9c9708917ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.342618] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294321, 'name': CreateVM_Task, 'duration_secs': 0.58162} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.343627] env[63021]: INFO nova.scheduler.client.report [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Deleted allocation for migration 863e91bb-e79b-4422-99bf-1ac76f99aafa [ 1090.344452] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1090.347420] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1090.347592] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1090.347912] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1090.348598] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0cca45da-c69a-486d-9442-6005138c9e40 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.353716] env[63021]: DEBUG oslo_vmware.api [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1090.353716] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52fbf7ed-b4af-f98c-8499-a79e76a36325" [ 1090.353716] env[63021]: _type = "Task" [ 1090.353716] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.364534] env[63021]: DEBUG oslo_vmware.api [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52fbf7ed-b4af-f98c-8499-a79e76a36325, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.427449] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.427714] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.429738] env[63021]: INFO nova.compute.claims [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1090.662770] env[63021]: DEBUG nova.objects.instance [None req-eb061d70-dbcb-4a53-8d2a-c0b6d094c646 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lazy-loading 'flavor' on Instance uuid 8d9387b8-6a84-4ada-8ac7-4ffd922b675a {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1090.666144] env[63021]: DEBUG nova.network.neutron [-] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.770773] env[63021]: DEBUG nova.network.neutron [-] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.797540] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1ee7b6f6-ce9d-4809-8dd1-57fe1e632182 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.807257] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ad0ce90-dc09-4fc4-b46a-9260febbdea1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.835591] env[63021]: DEBUG oslo_concurrency.lockutils [req-73144d8a-cb39-4707-990d-cf182ebd6cae req-ece7ccf9-7c0e-4870-8f84-85971e677c5b service nova] Releasing lock "refresh_cache-02c217fb-174a-47f8-a9b7-d71497797590" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1090.836045] env[63021]: DEBUG nova.compute.manager [req-f209ec01-eed0-40fe-8497-954f9842178f req-c9664a47-97e5-4dc4-b5a7-e4a3deb0eaf4 service nova] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Detach interface failed, port_id=905466dc-1047-469f-b282-768293b6d154, reason: Instance 418a394b-dbb2-47a2-96a9-a1caa4c4bb51 could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1090.850391] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "e940f625-f35e-459d-b8be-fa5916ce6efb" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.589s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.865545] env[63021]: DEBUG oslo_vmware.api [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52fbf7ed-b4af-f98c-8499-a79e76a36325, 'name': SearchDatastore_Task, 'duration_secs': 0.009672} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.865942] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1090.866272] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1090.866598] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1090.866811] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1090.867436] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1090.868576] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-558644cd-1014-4a7e-a07a-cf640cfce77b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.876687] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1090.876865] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1090.877923] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c0f4d9b-cc06-42e3-8a49-55e0f81a605d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.883082] env[63021]: DEBUG oslo_vmware.api [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1090.883082] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]526b7931-f06b-61f6-0a0d-05ff7bdeb72a" [ 1090.883082] env[63021]: _type = "Task" [ 1090.883082] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.890095] env[63021]: DEBUG oslo_vmware.api [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]526b7931-f06b-61f6-0a0d-05ff7bdeb72a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.168049] env[63021]: INFO nova.compute.manager [-] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Took 1.48 seconds to deallocate network for instance. [ 1091.273955] env[63021]: INFO nova.compute.manager [-] [instance: 418a394b-dbb2-47a2-96a9-a1caa4c4bb51] Took 1.34 seconds to deallocate network for instance. [ 1091.395926] env[63021]: DEBUG oslo_vmware.api [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]526b7931-f06b-61f6-0a0d-05ff7bdeb72a, 'name': SearchDatastore_Task, 'duration_secs': 0.009191} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.396731] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c438bfb0-0ca6-4f57-8e1f-7cbf7d677b05 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.401642] env[63021]: DEBUG oslo_vmware.api [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1091.401642] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52acada7-29db-18e5-a0ba-581351cf3eb7" [ 1091.401642] env[63021]: _type = "Task" [ 1091.401642] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.409339] env[63021]: DEBUG oslo_vmware.api [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52acada7-29db-18e5-a0ba-581351cf3eb7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.570264] env[63021]: DEBUG nova.compute.manager [req-ae3d789b-8628-4610-b8f9-817f3d2e7eeb req-9526a27c-91c9-4ca3-aae4-b1ce361d85c8 service nova] [instance: d32d9644-de8b-464e-ae16-7fa775a622ee] Received event network-vif-deleted-b7c97772-4a14-4bf5-9b07-980a1b3e92b6 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1091.572170] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abd3a1dc-1c74-4e3e-8cb9-3d9fe4bc6466 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.579480] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ffd432f-cba0-423b-b7c3-44b1b17010d8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.611711] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bf394c8-0968-4815-a0e6-cbd78c0ed1dd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.619013] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2489c2d5-ff85-4965-a392-70e96074049d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.632241] env[63021]: DEBUG nova.compute.provider_tree [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1091.671750] env[63021]: DEBUG oslo_concurrency.lockutils [None req-eb061d70-dbcb-4a53-8d2a-c0b6d094c646 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "8d9387b8-6a84-4ada-8ac7-4ffd922b675a" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.222s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.674697] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.780919] env[63021]: DEBUG oslo_concurrency.lockutils [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.873604] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquiring lock "8d9387b8-6a84-4ada-8ac7-4ffd922b675a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.873838] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "8d9387b8-6a84-4ada-8ac7-4ffd922b675a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.874067] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquiring lock "8d9387b8-6a84-4ada-8ac7-4ffd922b675a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.874260] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "8d9387b8-6a84-4ada-8ac7-4ffd922b675a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.874438] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "8d9387b8-6a84-4ada-8ac7-4ffd922b675a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.877942] env[63021]: INFO nova.compute.manager [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Terminating instance [ 1091.880066] env[63021]: DEBUG nova.compute.manager [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1091.881394] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1091.881394] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f72c3d32-a65d-410a-a8c1-43a197731a0c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.889282] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1091.889534] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-43d9a519-cdbb-43a0-ac4b-06a7f6372f06 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.896563] env[63021]: DEBUG oslo_vmware.api [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 1091.896563] env[63021]: value = "task-1294325" [ 1091.896563] env[63021]: _type = "Task" [ 1091.896563] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.905710] env[63021]: DEBUG oslo_vmware.api [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294325, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.915758] env[63021]: DEBUG oslo_vmware.api [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52acada7-29db-18e5-a0ba-581351cf3eb7, 'name': SearchDatastore_Task, 'duration_secs': 0.010217} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.916048] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1091.916407] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 02c217fb-174a-47f8-a9b7-d71497797590/02c217fb-174a-47f8-a9b7-d71497797590.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1091.916694] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0fdcb978-aa15-4a18-a266-86740d82667c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.923256] env[63021]: DEBUG oslo_vmware.api [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1091.923256] env[63021]: value = "task-1294326" [ 1091.923256] env[63021]: _type = "Task" [ 1091.923256] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.930892] env[63021]: DEBUG oslo_vmware.api [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294326, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.135111] env[63021]: DEBUG nova.scheduler.client.report [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1092.282458] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "e940f625-f35e-459d-b8be-fa5916ce6efb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.282811] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "e940f625-f35e-459d-b8be-fa5916ce6efb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.283035] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "e940f625-f35e-459d-b8be-fa5916ce6efb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.283258] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "e940f625-f35e-459d-b8be-fa5916ce6efb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.283508] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "e940f625-f35e-459d-b8be-fa5916ce6efb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.285988] env[63021]: INFO nova.compute.manager [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Terminating instance [ 1092.288343] env[63021]: DEBUG nova.compute.manager [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1092.288557] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1092.289482] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaf5ff9d-6232-4a68-ab15-e01b1af7b7ad {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.298522] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1092.298808] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3972def9-ef77-4eb6-a8e9-ff0e16a84c18 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.305183] env[63021]: DEBUG oslo_vmware.api [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1092.305183] env[63021]: value = "task-1294327" [ 1092.305183] env[63021]: _type = "Task" [ 1092.305183] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.316871] env[63021]: DEBUG oslo_vmware.api [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294327, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.406362] env[63021]: DEBUG oslo_vmware.api [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294325, 'name': PowerOffVM_Task, 'duration_secs': 0.199105} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.406613] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1092.406785] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1092.407054] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-40c682c5-b07f-490b-99a4-7f940aaf6990 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.432489] env[63021]: DEBUG oslo_vmware.api [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294326, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.478472} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.432737] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 02c217fb-174a-47f8-a9b7-d71497797590/02c217fb-174a-47f8-a9b7-d71497797590.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1092.432952] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1092.433209] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-80a74fbe-da85-45f8-8cb5-8845b9ca4b3c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.439564] env[63021]: DEBUG oslo_vmware.api [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1092.439564] env[63021]: value = "task-1294329" [ 1092.439564] env[63021]: _type = "Task" [ 1092.439564] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.447308] env[63021]: DEBUG oslo_vmware.api [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294329, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.475853] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1092.476102] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Deleting contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1092.476279] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Deleting the datastore file [datastore2] 8d9387b8-6a84-4ada-8ac7-4ffd922b675a {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1092.476550] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-da9bcfab-5777-4ce4-96fd-f80290d601de {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.482611] env[63021]: DEBUG oslo_vmware.api [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for the task: (returnval){ [ 1092.482611] env[63021]: value = "task-1294330" [ 1092.482611] env[63021]: _type = "Task" [ 1092.482611] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.491296] env[63021]: DEBUG oslo_vmware.api [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294330, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.600741] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1092.600741] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1092.600741] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1092.600969] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1092.600969] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63021) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1092.601083] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1092.601216] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Cleaning up deleted instances {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1092.639879] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.212s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.640249] env[63021]: DEBUG nova.compute.manager [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1092.642717] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.968s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.642940] env[63021]: DEBUG nova.objects.instance [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Lazy-loading 'resources' on Instance uuid d32d9644-de8b-464e-ae16-7fa775a622ee {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1092.814805] env[63021]: DEBUG oslo_vmware.api [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294327, 'name': PowerOffVM_Task, 'duration_secs': 0.235003} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.815341] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1092.815530] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1092.815796] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-59c2fa62-4830-4805-85e7-9178ce0cbfa6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.873677] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1092.873914] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1092.874118] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Deleting the datastore file [datastore1] e940f625-f35e-459d-b8be-fa5916ce6efb {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1092.874403] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8f2d724f-25d3-4271-8377-1d46c82f0971 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.880476] env[63021]: DEBUG oslo_vmware.api [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for the task: (returnval){ [ 1092.880476] env[63021]: value = "task-1294332" [ 1092.880476] env[63021]: _type = "Task" [ 1092.880476] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.888091] env[63021]: DEBUG oslo_vmware.api [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294332, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.949313] env[63021]: DEBUG oslo_vmware.api [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294329, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060867} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.949555] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1092.950396] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5460194a-dccf-428f-9be7-4e29f63cd8b3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.972728] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] 02c217fb-174a-47f8-a9b7-d71497797590/02c217fb-174a-47f8-a9b7-d71497797590.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1092.972968] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dc4eacd7-5c01-4059-9931-3d6624cc5cea {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.995793] env[63021]: DEBUG oslo_vmware.api [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Task: {'id': task-1294330, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.128014} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.996950] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1092.997166] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Deleted contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1092.997371] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1092.997558] env[63021]: INFO nova.compute.manager [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1092.997792] env[63021]: DEBUG oslo.service.loopingcall [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1092.998109] env[63021]: DEBUG oslo_vmware.api [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1092.998109] env[63021]: value = "task-1294333" [ 1092.998109] env[63021]: _type = "Task" [ 1092.998109] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.998278] env[63021]: DEBUG nova.compute.manager [-] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1092.998381] env[63021]: DEBUG nova.network.neutron [-] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1093.008566] env[63021]: DEBUG oslo_vmware.api [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294333, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.115448] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] There are 44 instances to clean {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1093.115592] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: a2d6db04-9293-4755-ad8c-2e46467f92cb] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1093.145675] env[63021]: DEBUG nova.compute.utils [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1093.146971] env[63021]: DEBUG nova.compute.manager [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1093.147151] env[63021]: DEBUG nova.network.neutron [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1093.212635] env[63021]: DEBUG nova.policy [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ba16721213dd49a3b4e4990d7d319d1f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3c8959d2f1fe4706b7407c11e128b773', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 1093.289146] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab4a45fd-c76b-405c-bed7-113ff5abe34f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.297395] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-532e720d-de70-40ed-b0dc-31140a93919d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.327761] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8da6f4bf-988d-4d83-87b0-3c16359aefde {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.335056] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-525c52ba-4e58-4cf0-9cb9-aea544468a5b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.348131] env[63021]: DEBUG nova.compute.provider_tree [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1093.389777] env[63021]: DEBUG oslo_vmware.api [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Task: {'id': task-1294332, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152361} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.390081] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1093.390227] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1093.390413] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1093.390588] env[63021]: INFO nova.compute.manager [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1093.390829] env[63021]: DEBUG oslo.service.loopingcall [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1093.391034] env[63021]: DEBUG nova.compute.manager [-] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1093.391131] env[63021]: DEBUG nova.network.neutron [-] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1093.511765] env[63021]: DEBUG oslo_vmware.api [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294333, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.619597] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: e2e98322-fdda-45e0-b3ba-58349ab940be] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1093.650919] env[63021]: DEBUG nova.compute.manager [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1093.692497] env[63021]: DEBUG nova.compute.manager [req-59381f19-4388-43dd-bbe1-b966150ed0da req-7c8d444d-f253-4168-94ae-10e9941e8c48 service nova] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Received event network-vif-deleted-d2876adc-ed3c-4540-b06d-fd179ecd2445 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1093.692682] env[63021]: INFO nova.compute.manager [req-59381f19-4388-43dd-bbe1-b966150ed0da req-7c8d444d-f253-4168-94ae-10e9941e8c48 service nova] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Neutron deleted interface d2876adc-ed3c-4540-b06d-fd179ecd2445; detaching it from the instance and deleting it from the info cache [ 1093.692861] env[63021]: DEBUG nova.network.neutron [req-59381f19-4388-43dd-bbe1-b966150ed0da req-7c8d444d-f253-4168-94ae-10e9941e8c48 service nova] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.812548] env[63021]: DEBUG nova.compute.manager [req-9a79c2ac-7fb2-4119-87e6-e970fa297b74 req-fdeac5d7-c6ff-4697-a7f5-3abb3864729a service nova] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Received event network-vif-deleted-1bb789b9-f77e-4e8c-a00b-c513af834d07 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1093.812754] env[63021]: INFO nova.compute.manager [req-9a79c2ac-7fb2-4119-87e6-e970fa297b74 req-fdeac5d7-c6ff-4697-a7f5-3abb3864729a service nova] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Neutron deleted interface 1bb789b9-f77e-4e8c-a00b-c513af834d07; detaching it from the instance and deleting it from the info cache [ 1093.812929] env[63021]: DEBUG nova.network.neutron [req-9a79c2ac-7fb2-4119-87e6-e970fa297b74 req-fdeac5d7-c6ff-4697-a7f5-3abb3864729a service nova] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.826273] env[63021]: DEBUG nova.network.neutron [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Successfully created port: 075306f8-9f3c-4d8a-9c1e-7422ef01d4c7 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1093.851889] env[63021]: DEBUG nova.scheduler.client.report [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1094.010675] env[63021]: DEBUG oslo_vmware.api [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294333, 'name': ReconfigVM_Task, 'duration_secs': 0.678291} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.010963] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Reconfigured VM instance instance-0000006a to attach disk [datastore1] 02c217fb-174a-47f8-a9b7-d71497797590/02c217fb-174a-47f8-a9b7-d71497797590.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1094.011643] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-20437517-2306-4ddf-bdda-a9b9ef64ecee {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.017742] env[63021]: DEBUG oslo_vmware.api [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1094.017742] env[63021]: value = "task-1294334" [ 1094.017742] env[63021]: _type = "Task" [ 1094.017742] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.026197] env[63021]: DEBUG oslo_vmware.api [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294334, 'name': Rename_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.122478] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: acc38ba0-515d-41d4-81dd-2b4374e033e7] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1094.126725] env[63021]: DEBUG nova.network.neutron [-] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.199018] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-279a1393-b3e2-45af-8457-2314879529ac {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.209701] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-370e35a4-34b9-4a31-b882-48512f517019 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.235895] env[63021]: DEBUG nova.compute.manager [req-59381f19-4388-43dd-bbe1-b966150ed0da req-7c8d444d-f253-4168-94ae-10e9941e8c48 service nova] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Detach interface failed, port_id=d2876adc-ed3c-4540-b06d-fd179ecd2445, reason: Instance e940f625-f35e-459d-b8be-fa5916ce6efb could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1094.292848] env[63021]: DEBUG nova.network.neutron [-] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.315737] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f416a045-198a-44a2-9dd9-3d01bd719e58 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.327723] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a183f7c-3983-47ef-b265-210e198aaa43 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.353923] env[63021]: DEBUG nova.compute.manager [req-9a79c2ac-7fb2-4119-87e6-e970fa297b74 req-fdeac5d7-c6ff-4697-a7f5-3abb3864729a service nova] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Detach interface failed, port_id=1bb789b9-f77e-4e8c-a00b-c513af834d07, reason: Instance 8d9387b8-6a84-4ada-8ac7-4ffd922b675a could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1094.356904] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.714s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.359161] env[63021]: DEBUG oslo_concurrency.lockutils [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.578s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1094.361139] env[63021]: DEBUG nova.objects.instance [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lazy-loading 'resources' on Instance uuid 418a394b-dbb2-47a2-96a9-a1caa4c4bb51 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1094.386063] env[63021]: INFO nova.scheduler.client.report [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Deleted allocations for instance d32d9644-de8b-464e-ae16-7fa775a622ee [ 1094.529073] env[63021]: DEBUG oslo_vmware.api [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294334, 'name': Rename_Task, 'duration_secs': 0.141916} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.529328] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1094.529651] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fb5bb93c-6632-4b9a-a5cc-a828ff7defb8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.535368] env[63021]: DEBUG oslo_vmware.api [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1094.535368] env[63021]: value = "task-1294335" [ 1094.535368] env[63021]: _type = "Task" [ 1094.535368] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.543271] env[63021]: DEBUG oslo_vmware.api [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294335, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.625614] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 63d5b0e4-0d61-4fa7-b193-2eaa171fc461] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1094.630062] env[63021]: INFO nova.compute.manager [-] [instance: e940f625-f35e-459d-b8be-fa5916ce6efb] Took 1.24 seconds to deallocate network for instance. [ 1094.662347] env[63021]: DEBUG nova.compute.manager [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1094.687904] env[63021]: DEBUG nova.virt.hardware [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1094.688188] env[63021]: DEBUG nova.virt.hardware [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1094.688356] env[63021]: DEBUG nova.virt.hardware [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1094.688537] env[63021]: DEBUG nova.virt.hardware [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1094.688685] env[63021]: DEBUG nova.virt.hardware [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1094.688852] env[63021]: DEBUG nova.virt.hardware [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1094.689053] env[63021]: DEBUG nova.virt.hardware [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1094.689247] env[63021]: DEBUG nova.virt.hardware [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1094.689446] env[63021]: DEBUG nova.virt.hardware [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1094.689641] env[63021]: DEBUG nova.virt.hardware [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1094.689831] env[63021]: DEBUG nova.virt.hardware [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1094.690936] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01b54740-0f20-4320-9f9d-2783b676df59 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.698633] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44767c7f-ba29-4180-a02b-db1761490f20 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.793151] env[63021]: INFO nova.compute.manager [-] [instance: 8d9387b8-6a84-4ada-8ac7-4ffd922b675a] Took 1.79 seconds to deallocate network for instance. [ 1094.896723] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f75f21cb-59c6-4344-a94d-9daf6318946d tempest-ServersTestFqdnHostnames-189775116 tempest-ServersTestFqdnHostnames-189775116-project-member] Lock "d32d9644-de8b-464e-ae16-7fa775a622ee" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.331s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.996241] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6142bf0c-b1de-4045-8986-f826ea7e5513 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.005408] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0525392-27ea-4896-aa38-fba47b33022b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.042370] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f75df6c-6740-41ba-869f-99695decbed0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.052133] env[63021]: DEBUG oslo_vmware.api [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294335, 'name': PowerOnVM_Task, 'duration_secs': 0.427174} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.053384] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cce612b4-708c-48dc-8763-2eb069fb79d7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.057117] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1095.057392] env[63021]: INFO nova.compute.manager [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Took 7.82 seconds to spawn the instance on the hypervisor. [ 1095.057612] env[63021]: DEBUG nova.compute.manager [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1095.059039] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45e9b691-1c13-45fd-8f22-a3ce1f6d5e3d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.070602] env[63021]: DEBUG nova.compute.provider_tree [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1095.129092] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 07510760-13ff-4797-86ca-86b384aacdfa] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1095.136573] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.300056] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.318709] env[63021]: DEBUG nova.compute.manager [req-950b1778-f935-4f11-8c55-e7da21a4079e req-72a41034-ab30-47e4-8a25-e09a115442aa service nova] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Received event network-vif-plugged-075306f8-9f3c-4d8a-9c1e-7422ef01d4c7 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1095.318978] env[63021]: DEBUG oslo_concurrency.lockutils [req-950b1778-f935-4f11-8c55-e7da21a4079e req-72a41034-ab30-47e4-8a25-e09a115442aa service nova] Acquiring lock "ae8a512d-5393-432c-bca0-4dc86711891b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.319248] env[63021]: DEBUG oslo_concurrency.lockutils [req-950b1778-f935-4f11-8c55-e7da21a4079e req-72a41034-ab30-47e4-8a25-e09a115442aa service nova] Lock "ae8a512d-5393-432c-bca0-4dc86711891b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.319465] env[63021]: DEBUG oslo_concurrency.lockutils [req-950b1778-f935-4f11-8c55-e7da21a4079e req-72a41034-ab30-47e4-8a25-e09a115442aa service nova] Lock "ae8a512d-5393-432c-bca0-4dc86711891b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.319671] env[63021]: DEBUG nova.compute.manager [req-950b1778-f935-4f11-8c55-e7da21a4079e req-72a41034-ab30-47e4-8a25-e09a115442aa service nova] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] No waiting events found dispatching network-vif-plugged-075306f8-9f3c-4d8a-9c1e-7422ef01d4c7 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1095.319871] env[63021]: WARNING nova.compute.manager [req-950b1778-f935-4f11-8c55-e7da21a4079e req-72a41034-ab30-47e4-8a25-e09a115442aa service nova] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Received unexpected event network-vif-plugged-075306f8-9f3c-4d8a-9c1e-7422ef01d4c7 for instance with vm_state building and task_state spawning. [ 1095.577401] env[63021]: DEBUG nova.scheduler.client.report [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1095.587559] env[63021]: INFO nova.compute.manager [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Took 12.66 seconds to build instance. [ 1095.632123] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 023c0891-2b08-48bb-9754-408ce7d04c9c] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1095.916107] env[63021]: DEBUG nova.network.neutron [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Successfully updated port: 075306f8-9f3c-4d8a-9c1e-7422ef01d4c7 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1095.952405] env[63021]: DEBUG nova.compute.manager [req-b20910ab-e3d7-4c5d-9895-b91edab7e98a req-3e794404-c842-4a66-87d3-84205741c22c service nova] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Received event network-changed-075306f8-9f3c-4d8a-9c1e-7422ef01d4c7 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1095.952695] env[63021]: DEBUG nova.compute.manager [req-b20910ab-e3d7-4c5d-9895-b91edab7e98a req-3e794404-c842-4a66-87d3-84205741c22c service nova] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Refreshing instance network info cache due to event network-changed-075306f8-9f3c-4d8a-9c1e-7422ef01d4c7. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1095.952951] env[63021]: DEBUG oslo_concurrency.lockutils [req-b20910ab-e3d7-4c5d-9895-b91edab7e98a req-3e794404-c842-4a66-87d3-84205741c22c service nova] Acquiring lock "refresh_cache-ae8a512d-5393-432c-bca0-4dc86711891b" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1095.953117] env[63021]: DEBUG oslo_concurrency.lockutils [req-b20910ab-e3d7-4c5d-9895-b91edab7e98a req-3e794404-c842-4a66-87d3-84205741c22c service nova] Acquired lock "refresh_cache-ae8a512d-5393-432c-bca0-4dc86711891b" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.953285] env[63021]: DEBUG nova.network.neutron [req-b20910ab-e3d7-4c5d-9895-b91edab7e98a req-3e794404-c842-4a66-87d3-84205741c22c service nova] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Refreshing network info cache for port 075306f8-9f3c-4d8a-9c1e-7422ef01d4c7 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1096.088332] env[63021]: DEBUG oslo_concurrency.lockutils [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.729s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.092513] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.956s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.092678] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.095598] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.796s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.097838] env[63021]: DEBUG nova.objects.instance [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lazy-loading 'resources' on Instance uuid 8d9387b8-6a84-4ada-8ac7-4ffd922b675a {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1096.097838] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0cb0282b-2dee-487e-bca6-654f19fb3f51 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "02c217fb-174a-47f8-a9b7-d71497797590" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.176s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.121068] env[63021]: INFO nova.scheduler.client.report [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Deleted allocations for instance e940f625-f35e-459d-b8be-fa5916ce6efb [ 1096.126702] env[63021]: INFO nova.scheduler.client.report [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Deleted allocations for instance 418a394b-dbb2-47a2-96a9-a1caa4c4bb51 [ 1096.138822] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 67aed8eb-d08c-47cc-8558-5c5a0f058d6e] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1096.292463] env[63021]: DEBUG nova.compute.manager [req-adcaef62-4c70-4d86-8511-7b74322531d1 req-b712fdb9-51f5-4586-bc81-45f5fd537ff3 service nova] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Received event network-changed-416fe99a-8969-4bb8-90d8-b9c9708917ca {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1096.292463] env[63021]: DEBUG nova.compute.manager [req-adcaef62-4c70-4d86-8511-7b74322531d1 req-b712fdb9-51f5-4586-bc81-45f5fd537ff3 service nova] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Refreshing instance network info cache due to event network-changed-416fe99a-8969-4bb8-90d8-b9c9708917ca. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1096.292463] env[63021]: DEBUG oslo_concurrency.lockutils [req-adcaef62-4c70-4d86-8511-7b74322531d1 req-b712fdb9-51f5-4586-bc81-45f5fd537ff3 service nova] Acquiring lock "refresh_cache-02c217fb-174a-47f8-a9b7-d71497797590" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.292463] env[63021]: DEBUG oslo_concurrency.lockutils [req-adcaef62-4c70-4d86-8511-7b74322531d1 req-b712fdb9-51f5-4586-bc81-45f5fd537ff3 service nova] Acquired lock "refresh_cache-02c217fb-174a-47f8-a9b7-d71497797590" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.292618] env[63021]: DEBUG nova.network.neutron [req-adcaef62-4c70-4d86-8511-7b74322531d1 req-b712fdb9-51f5-4586-bc81-45f5fd537ff3 service nova] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Refreshing network info cache for port 416fe99a-8969-4bb8-90d8-b9c9708917ca {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1096.424635] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "refresh_cache-ae8a512d-5393-432c-bca0-4dc86711891b" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.484782] env[63021]: DEBUG nova.network.neutron [req-b20910ab-e3d7-4c5d-9895-b91edab7e98a req-3e794404-c842-4a66-87d3-84205741c22c service nova] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1096.573495] env[63021]: DEBUG nova.network.neutron [req-b20910ab-e3d7-4c5d-9895-b91edab7e98a req-3e794404-c842-4a66-87d3-84205741c22c service nova] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1096.639899] env[63021]: DEBUG oslo_concurrency.lockutils [None req-04331e13-8e29-41da-a668-72f5130cc19d tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "418a394b-dbb2-47a2-96a9-a1caa4c4bb51" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.864s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.641096] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 9f822151-3307-4093-a972-8af128ea1892] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1096.643778] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2249de97-81f6-44ee-a115-7574dd48678b tempest-DeleteServersTestJSON-492218925 tempest-DeleteServersTestJSON-492218925-project-member] Lock "e940f625-f35e-459d-b8be-fa5916ce6efb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.361s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.729380] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ac55424-e334-41bb-b619-7fd8ad69cb61 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.737327] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68cbe77a-f18d-4784-b7b3-9bbc4991947a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.768656] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6281360-341d-450b-9b05-4dce0b05a78c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.776174] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2375c520-a06f-4206-ae17-dedaa8d4e154 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.791959] env[63021]: DEBUG nova.compute.provider_tree [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1097.037326] env[63021]: DEBUG nova.network.neutron [req-adcaef62-4c70-4d86-8511-7b74322531d1 req-b712fdb9-51f5-4586-bc81-45f5fd537ff3 service nova] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Updated VIF entry in instance network info cache for port 416fe99a-8969-4bb8-90d8-b9c9708917ca. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1097.037544] env[63021]: DEBUG nova.network.neutron [req-adcaef62-4c70-4d86-8511-7b74322531d1 req-b712fdb9-51f5-4586-bc81-45f5fd537ff3 service nova] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Updating instance_info_cache with network_info: [{"id": "416fe99a-8969-4bb8-90d8-b9c9708917ca", "address": "fa:16:3e:71:2a:8b", "network": {"id": "d27bb3cc-fb80-4b03-9288-ed17a0d05140", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-122910456-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "547d9a3099b94ff488ef6929ae591bac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1c797172-a569-458e-aeb0-3f21e589a740", "external-id": "nsx-vlan-transportzone-957", "segmentation_id": 957, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap416fe99a-89", "ovs_interfaceid": "416fe99a-8969-4bb8-90d8-b9c9708917ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1097.075649] env[63021]: DEBUG oslo_concurrency.lockutils [req-b20910ab-e3d7-4c5d-9895-b91edab7e98a req-3e794404-c842-4a66-87d3-84205741c22c service nova] Releasing lock "refresh_cache-ae8a512d-5393-432c-bca0-4dc86711891b" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1097.076038] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquired lock "refresh_cache-ae8a512d-5393-432c-bca0-4dc86711891b" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.076206] env[63021]: DEBUG nova.network.neutron [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1097.145366] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 67f736d9-82ec-448e-8b72-84fdd704aa8d] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1097.294883] env[63021]: DEBUG nova.scheduler.client.report [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1097.492993] env[63021]: DEBUG oslo_concurrency.lockutils [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "ae103118-bddf-46fe-90b9-98b60952ebba" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.493387] env[63021]: DEBUG oslo_concurrency.lockutils [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "ae103118-bddf-46fe-90b9-98b60952ebba" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.493669] env[63021]: DEBUG oslo_concurrency.lockutils [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "ae103118-bddf-46fe-90b9-98b60952ebba-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.494030] env[63021]: DEBUG oslo_concurrency.lockutils [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "ae103118-bddf-46fe-90b9-98b60952ebba-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.494188] env[63021]: DEBUG oslo_concurrency.lockutils [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "ae103118-bddf-46fe-90b9-98b60952ebba-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.496587] env[63021]: INFO nova.compute.manager [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Terminating instance [ 1097.498576] env[63021]: DEBUG nova.compute.manager [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1097.498779] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1097.499657] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-120af74c-bdf4-4f32-9cf2-baccbcd48a19 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.507964] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1097.508230] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5792e2b5-d75f-478e-a66a-fc64ef49c4b3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.514907] env[63021]: DEBUG oslo_vmware.api [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 1097.514907] env[63021]: value = "task-1294337" [ 1097.514907] env[63021]: _type = "Task" [ 1097.514907] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.522979] env[63021]: DEBUG oslo_vmware.api [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294337, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.539803] env[63021]: DEBUG oslo_concurrency.lockutils [req-adcaef62-4c70-4d86-8511-7b74322531d1 req-b712fdb9-51f5-4586-bc81-45f5fd537ff3 service nova] Releasing lock "refresh_cache-02c217fb-174a-47f8-a9b7-d71497797590" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1097.609699] env[63021]: DEBUG nova.network.neutron [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1097.648842] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 7df37eb7-ffc1-4b8e-9aba-caaf3c5338df] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1097.757984] env[63021]: DEBUG nova.network.neutron [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Updating instance_info_cache with network_info: [{"id": "075306f8-9f3c-4d8a-9c1e-7422ef01d4c7", "address": "fa:16:3e:05:af:b7", "network": {"id": "0c0dd1a7-6687-45c5-82b1-33f34bdbb848", "bridge": "br-int", "label": "tempest-ServersTestJSON-523240467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c8959d2f1fe4706b7407c11e128b773", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffc811e4-0e4e-4748-8c7e-b3f14ccbd42d", "external-id": "nsx-vlan-transportzone-404", "segmentation_id": 404, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap075306f8-9f", "ovs_interfaceid": "075306f8-9f3c-4d8a-9c1e-7422ef01d4c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1097.799596] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.704s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.821082] env[63021]: INFO nova.scheduler.client.report [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Deleted allocations for instance 8d9387b8-6a84-4ada-8ac7-4ffd922b675a [ 1098.025360] env[63021]: DEBUG oslo_vmware.api [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294337, 'name': PowerOffVM_Task, 'duration_secs': 0.197225} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.025645] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1098.025843] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1098.026079] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-52ca8815-92b3-4313-85bd-d5208ac3a01a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.089224] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1098.089325] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1098.089549] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Deleting the datastore file [datastore1] ae103118-bddf-46fe-90b9-98b60952ebba {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1098.089840] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-020547bf-2f46-4483-8847-ff3b81320deb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.096303] env[63021]: DEBUG oslo_vmware.api [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for the task: (returnval){ [ 1098.096303] env[63021]: value = "task-1294339" [ 1098.096303] env[63021]: _type = "Task" [ 1098.096303] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.104662] env[63021]: DEBUG oslo_vmware.api [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294339, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.152502] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 6155c098-e4a6-47e6-b343-4a77ca90eb2e] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1098.259934] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Releasing lock "refresh_cache-ae8a512d-5393-432c-bca0-4dc86711891b" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1098.260292] env[63021]: DEBUG nova.compute.manager [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Instance network_info: |[{"id": "075306f8-9f3c-4d8a-9c1e-7422ef01d4c7", "address": "fa:16:3e:05:af:b7", "network": {"id": "0c0dd1a7-6687-45c5-82b1-33f34bdbb848", "bridge": "br-int", "label": "tempest-ServersTestJSON-523240467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c8959d2f1fe4706b7407c11e128b773", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffc811e4-0e4e-4748-8c7e-b3f14ccbd42d", "external-id": "nsx-vlan-transportzone-404", "segmentation_id": 404, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap075306f8-9f", "ovs_interfaceid": "075306f8-9f3c-4d8a-9c1e-7422ef01d4c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1098.260737] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:05:af:b7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ffc811e4-0e4e-4748-8c7e-b3f14ccbd42d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '075306f8-9f3c-4d8a-9c1e-7422ef01d4c7', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1098.268362] env[63021]: DEBUG oslo.service.loopingcall [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1098.268620] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1098.268858] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6228a812-9cf2-4a23-8d9e-f928a38478bc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.289142] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1098.289142] env[63021]: value = "task-1294340" [ 1098.289142] env[63021]: _type = "Task" [ 1098.289142] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.296675] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294340, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.328485] env[63021]: DEBUG oslo_concurrency.lockutils [None req-0694235b-9115-4496-8e7e-5d2934ca7e32 tempest-AttachVolumeTestJSON-1066430688 tempest-AttachVolumeTestJSON-1066430688-project-member] Lock "8d9387b8-6a84-4ada-8ac7-4ffd922b675a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.455s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.606894] env[63021]: DEBUG oslo_vmware.api [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Task: {'id': task-1294339, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169386} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.607281] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1098.607423] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1098.607571] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1098.607770] env[63021]: INFO nova.compute.manager [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1098.608032] env[63021]: DEBUG oslo.service.loopingcall [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1098.608233] env[63021]: DEBUG nova.compute.manager [-] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1098.608346] env[63021]: DEBUG nova.network.neutron [-] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1098.655950] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 919887c8-171d-48fe-8c6b-2abf5f55d6ab] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1098.799353] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294340, 'name': CreateVM_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.887364] env[63021]: DEBUG nova.compute.manager [req-f7d9500f-5451-47d7-96ad-e9ce3e371bb5 req-56b482cb-20f2-4561-bc9c-0be26cfab933 service nova] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Received event network-vif-deleted-f99a201d-df02-499f-85d9-2566ac1dd82b {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1098.887674] env[63021]: INFO nova.compute.manager [req-f7d9500f-5451-47d7-96ad-e9ce3e371bb5 req-56b482cb-20f2-4561-bc9c-0be26cfab933 service nova] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Neutron deleted interface f99a201d-df02-499f-85d9-2566ac1dd82b; detaching it from the instance and deleting it from the info cache [ 1098.887911] env[63021]: DEBUG nova.network.neutron [req-f7d9500f-5451-47d7-96ad-e9ce3e371bb5 req-56b482cb-20f2-4561-bc9c-0be26cfab933 service nova] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1099.159700] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: a52e4cb5-d0ea-4698-9955-753626762a02] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1099.301200] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294340, 'name': CreateVM_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.367102] env[63021]: DEBUG nova.network.neutron [-] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1099.390751] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e26875e2-cb0c-4149-bd87-9cc42b9c8994 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.400710] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80ef6589-e610-4903-aaaf-73be1b87172b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.426235] env[63021]: DEBUG nova.compute.manager [req-f7d9500f-5451-47d7-96ad-e9ce3e371bb5 req-56b482cb-20f2-4561-bc9c-0be26cfab933 service nova] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Detach interface failed, port_id=f99a201d-df02-499f-85d9-2566ac1dd82b, reason: Instance ae103118-bddf-46fe-90b9-98b60952ebba could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1099.662793] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 207c8197-6191-4837-8ead-8c24aa5b35bc] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1099.803099] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294340, 'name': CreateVM_Task} progress is 99%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.870328] env[63021]: INFO nova.compute.manager [-] [instance: ae103118-bddf-46fe-90b9-98b60952ebba] Took 1.26 seconds to deallocate network for instance. [ 1100.166268] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 2d647028-306a-41d5-96f0-2e82a98ab56d] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1100.301932] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294340, 'name': CreateVM_Task, 'duration_secs': 1.580946} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.302126] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1100.302784] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1100.302955] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1100.303310] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1100.303578] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0bf3113a-44df-4a0f-9e01-24b56c1b5e3c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.308321] env[63021]: DEBUG oslo_vmware.api [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1100.308321] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52887794-f61e-860b-06cc-8953fc7e9082" [ 1100.308321] env[63021]: _type = "Task" [ 1100.308321] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.315850] env[63021]: DEBUG oslo_vmware.api [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52887794-f61e-860b-06cc-8953fc7e9082, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.376534] env[63021]: DEBUG oslo_concurrency.lockutils [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.376903] env[63021]: DEBUG oslo_concurrency.lockutils [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.377228] env[63021]: DEBUG nova.objects.instance [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lazy-loading 'resources' on Instance uuid ae103118-bddf-46fe-90b9-98b60952ebba {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1100.669600] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 140184cd-6e4c-425c-8b17-361a1f565e93] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1100.820014] env[63021]: DEBUG oslo_vmware.api [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52887794-f61e-860b-06cc-8953fc7e9082, 'name': SearchDatastore_Task, 'duration_secs': 0.010431} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.820430] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1100.820862] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1100.820985] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1100.821181] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1100.821399] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1100.821790] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-519245c8-3e04-4bbe-9940-1dcaa971b88b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.830500] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1100.830720] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1100.831461] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92b0ae9d-4d15-439b-85f7-ac120084b696 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.836772] env[63021]: DEBUG oslo_vmware.api [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1100.836772] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52e78f10-d56d-ccb6-7929-cb667924361d" [ 1100.836772] env[63021]: _type = "Task" [ 1100.836772] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.844417] env[63021]: DEBUG oslo_vmware.api [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52e78f10-d56d-ccb6-7929-cb667924361d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.082920] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe31f1c-6602-4473-8d17-82b0978ba8e2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.090896] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4661ef79-6451-4f4a-9207-a596227543ef {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.121922] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bf580a8-2efa-4d7d-a3ee-4f718b6c5162 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.130479] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21d06333-ac55-41d4-a18f-c040a7042282 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.142407] env[63021]: DEBUG nova.compute.provider_tree [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1101.173853] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: d63fd465-975f-42ce-b1be-ac8a6929bc4d] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1101.346905] env[63021]: DEBUG oslo_vmware.api [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52e78f10-d56d-ccb6-7929-cb667924361d, 'name': SearchDatastore_Task, 'duration_secs': 0.008778} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.347961] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ec07704-c4fc-4f50-b941-00fbe97f11c6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.352993] env[63021]: DEBUG oslo_vmware.api [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1101.352993] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5236ecc0-e740-3a87-256b-d5acbb34faf9" [ 1101.352993] env[63021]: _type = "Task" [ 1101.352993] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.361311] env[63021]: DEBUG oslo_vmware.api [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5236ecc0-e740-3a87-256b-d5acbb34faf9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.645662] env[63021]: DEBUG nova.scheduler.client.report [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1101.676355] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: ed26dff1-fed0-4baf-ad41-d14850254aec] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1101.863372] env[63021]: DEBUG oslo_vmware.api [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5236ecc0-e740-3a87-256b-d5acbb34faf9, 'name': SearchDatastore_Task, 'duration_secs': 0.01079} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.863664] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1101.863921] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] ae8a512d-5393-432c-bca0-4dc86711891b/ae8a512d-5393-432c-bca0-4dc86711891b.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1101.864202] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-55bbc28d-bb67-4988-b7d8-8dc06a8bd58a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.870578] env[63021]: DEBUG oslo_vmware.api [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1101.870578] env[63021]: value = "task-1294343" [ 1101.870578] env[63021]: _type = "Task" [ 1101.870578] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.877651] env[63021]: DEBUG oslo_vmware.api [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294343, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.150961] env[63021]: DEBUG oslo_concurrency.lockutils [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.774s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.179803] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 4d712cc1-6202-407e-8767-c620a1ce0bba] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1102.184061] env[63021]: INFO nova.scheduler.client.report [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Deleted allocations for instance ae103118-bddf-46fe-90b9-98b60952ebba [ 1102.380181] env[63021]: DEBUG oslo_vmware.api [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294343, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.466221} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.380484] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] ae8a512d-5393-432c-bca0-4dc86711891b/ae8a512d-5393-432c-bca0-4dc86711891b.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1102.380701] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1102.380950] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-be5de877-bb85-48d5-aa40-5029452882f6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.386615] env[63021]: DEBUG oslo_vmware.api [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1102.386615] env[63021]: value = "task-1294344" [ 1102.386615] env[63021]: _type = "Task" [ 1102.386615] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.393909] env[63021]: DEBUG oslo_vmware.api [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294344, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.686229] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 80b857c4-d9dd-4483-970f-a0f5f4bcf173] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1102.691050] env[63021]: DEBUG oslo_concurrency.lockutils [None req-077974d3-fdc2-42be-ae12-be49b63b24f2 tempest-ServerRescueNegativeTestJSON-41098401 tempest-ServerRescueNegativeTestJSON-41098401-project-member] Lock "ae103118-bddf-46fe-90b9-98b60952ebba" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.197s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.898988] env[63021]: DEBUG oslo_vmware.api [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294344, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067453} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.898988] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1102.899193] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eb37340-96d4-421f-928c-3eb820249be3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.926731] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] ae8a512d-5393-432c-bca0-4dc86711891b/ae8a512d-5393-432c-bca0-4dc86711891b.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1102.927071] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e9485aa-db1c-4b7f-bfcc-71cfda4b3164 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.946651] env[63021]: DEBUG oslo_vmware.api [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1102.946651] env[63021]: value = "task-1294345" [ 1102.946651] env[63021]: _type = "Task" [ 1102.946651] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.954941] env[63021]: DEBUG oslo_vmware.api [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294345, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.188933] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 1b4871cc-3dac-4683-8fce-93c4d9e4407d] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1103.456954] env[63021]: DEBUG oslo_vmware.api [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294345, 'name': ReconfigVM_Task, 'duration_secs': 0.29639} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.457337] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Reconfigured VM instance instance-0000006b to attach disk [datastore1] ae8a512d-5393-432c-bca0-4dc86711891b/ae8a512d-5393-432c-bca0-4dc86711891b.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1103.457957] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-734b7c11-49c6-42ad-956c-2179e8d55e1c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.464432] env[63021]: DEBUG oslo_vmware.api [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1103.464432] env[63021]: value = "task-1294346" [ 1103.464432] env[63021]: _type = "Task" [ 1103.464432] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.472858] env[63021]: DEBUG oslo_vmware.api [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294346, 'name': Rename_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.692494] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 51e4a14c-4792-43cd-9e99-4e06a8f6d6fb] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1103.975768] env[63021]: DEBUG oslo_vmware.api [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294346, 'name': Rename_Task, 'duration_secs': 0.138045} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.976069] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1103.976330] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-92e6565b-cb73-4ac1-aa0f-3ab4c2a713e5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.982901] env[63021]: DEBUG oslo_vmware.api [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1103.982901] env[63021]: value = "task-1294347" [ 1103.982901] env[63021]: _type = "Task" [ 1103.982901] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.990759] env[63021]: DEBUG oslo_vmware.api [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294347, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.196028] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 50a1269e-4d83-4cf7-ba14-3b1afb3c43ed] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1104.492625] env[63021]: DEBUG oslo_vmware.api [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294347, 'name': PowerOnVM_Task} progress is 78%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.700084] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 44dadf8e-a727-448f-887d-6408f7ebfbc6] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1104.993147] env[63021]: DEBUG oslo_vmware.api [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294347, 'name': PowerOnVM_Task, 'duration_secs': 0.82783} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.993437] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1104.993650] env[63021]: INFO nova.compute.manager [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Took 10.33 seconds to spawn the instance on the hypervisor. [ 1104.993833] env[63021]: DEBUG nova.compute.manager [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1104.994635] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccc229ad-22dc-435a-91d7-873c7de1bc23 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.204018] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: f0154308-e6a8-4321-a082-99291344664e] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1105.513098] env[63021]: INFO nova.compute.manager [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Took 15.10 seconds to build instance. [ 1105.707493] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 881c2fb8-5f8c-48c2-a173-8fab949f24f7] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1106.014652] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ad689884-10ae-4eeb-931c-50869f724973 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "ae8a512d-5393-432c-bca0-4dc86711891b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.613s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1106.211069] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 16b21798-17e8-4aeb-affa-57eae31d5dd3] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1106.574681] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "ae8a512d-5393-432c-bca0-4dc86711891b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.574971] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "ae8a512d-5393-432c-bca0-4dc86711891b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.575688] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "ae8a512d-5393-432c-bca0-4dc86711891b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.575912] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "ae8a512d-5393-432c-bca0-4dc86711891b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.576114] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "ae8a512d-5393-432c-bca0-4dc86711891b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1106.578350] env[63021]: INFO nova.compute.manager [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Terminating instance [ 1106.580285] env[63021]: DEBUG nova.compute.manager [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1106.580480] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1106.581310] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfba06f7-c58a-42eb-a297-157510a7594c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.588993] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1106.589247] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f6cb8ee9-2b49-43d8-9b06-38d8345b87f7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.594855] env[63021]: DEBUG oslo_vmware.api [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1106.594855] env[63021]: value = "task-1294349" [ 1106.594855] env[63021]: _type = "Task" [ 1106.594855] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.602525] env[63021]: DEBUG oslo_vmware.api [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294349, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.714319] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: c328d2f7-3398-4f25-b11c-f464be7af8a1] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1107.105718] env[63021]: DEBUG oslo_vmware.api [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294349, 'name': PowerOffVM_Task, 'duration_secs': 0.223268} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.106157] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1107.106464] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1107.106929] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e53c7167-038d-4f0f-bae2-0e7725f9aa91 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.175235] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1107.175499] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1107.175758] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Deleting the datastore file [datastore1] ae8a512d-5393-432c-bca0-4dc86711891b {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1107.176091] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b866cbe0-8521-4af5-8696-664b20767e45 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.182670] env[63021]: DEBUG oslo_vmware.api [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1107.182670] env[63021]: value = "task-1294351" [ 1107.182670] env[63021]: _type = "Task" [ 1107.182670] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.190888] env[63021]: DEBUG oslo_vmware.api [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294351, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.218315] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 45c9e2be-eb19-4b83-b280-c9eeaddfccba] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1107.692584] env[63021]: DEBUG oslo_vmware.api [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294351, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.132391} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.692838] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1107.693030] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1107.693232] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1107.693443] env[63021]: INFO nova.compute.manager [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1107.693718] env[63021]: DEBUG oslo.service.loopingcall [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1107.693918] env[63021]: DEBUG nova.compute.manager [-] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1107.694018] env[63021]: DEBUG nova.network.neutron [-] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1107.722219] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: df856146-61ba-4de7-a5fd-d75263927163] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1108.094490] env[63021]: DEBUG nova.compute.manager [req-87515d9f-61ec-4f25-82b7-0aa787708be0 req-9232211f-ebe8-4dc1-b8e3-239e675a3ed7 service nova] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Received event network-vif-deleted-075306f8-9f3c-4d8a-9c1e-7422ef01d4c7 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1108.095072] env[63021]: INFO nova.compute.manager [req-87515d9f-61ec-4f25-82b7-0aa787708be0 req-9232211f-ebe8-4dc1-b8e3-239e675a3ed7 service nova] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Neutron deleted interface 075306f8-9f3c-4d8a-9c1e-7422ef01d4c7; detaching it from the instance and deleting it from the info cache [ 1108.095072] env[63021]: DEBUG nova.network.neutron [req-87515d9f-61ec-4f25-82b7-0aa787708be0 req-9232211f-ebe8-4dc1-b8e3-239e675a3ed7 service nova] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1108.225273] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 71909213-258d-45f5-9a3e-e473e8ce8aa2] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1108.568648] env[63021]: DEBUG nova.network.neutron [-] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1108.599027] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c4534718-464c-47b5-8523-0e765b019200 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.608192] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f09b994d-e93b-4137-ac75-b593ce976e22 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.632298] env[63021]: DEBUG nova.compute.manager [req-87515d9f-61ec-4f25-82b7-0aa787708be0 req-9232211f-ebe8-4dc1-b8e3-239e675a3ed7 service nova] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Detach interface failed, port_id=075306f8-9f3c-4d8a-9c1e-7422ef01d4c7, reason: Instance ae8a512d-5393-432c-bca0-4dc86711891b could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1108.728224] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 7787f6bb-e855-476f-9146-2a648932429d] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1109.071716] env[63021]: INFO nova.compute.manager [-] [instance: ae8a512d-5393-432c-bca0-4dc86711891b] Took 1.38 seconds to deallocate network for instance. [ 1109.232114] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 70f67905-7c60-433d-9ebc-d66fa44eb36e] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1109.578029] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.578335] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.578558] env[63021]: DEBUG nova.objects.instance [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lazy-loading 'resources' on Instance uuid ae8a512d-5393-432c-bca0-4dc86711891b {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1109.736511] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 6f4e6245-c0d9-46b4-baf5-5aca19e66da5] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1110.157842] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39838338-2b92-4f70-adc8-88350c6f7014 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.165536] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c195d67-f2e2-40f5-bfca-0e313c227afc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.196073] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26efa65d-d255-45b0-8742-d98576f25ae0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.203230] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-947b1895-c364-4f3b-9888-744a370c62b7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.216626] env[63021]: DEBUG nova.compute.provider_tree [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Updating inventory in ProviderTree for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1110.238078] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 2ed2b333-78a2-4ddd-ade4-4a1bcff8c6c4] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1110.736337] env[63021]: ERROR nova.scheduler.client.report [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [req-40a08250-0e30-43b5-b1a5-ba004f2a9f7c] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 6ef6c041-2b0d-4bc3-8662-5d679a77a86d. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-40a08250-0e30-43b5-b1a5-ba004f2a9f7c"}]} [ 1110.741019] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 36d26354-6127-4557-acb0-8743e27ff1ff] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1110.752404] env[63021]: DEBUG nova.scheduler.client.report [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Refreshing inventories for resource provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1110.765013] env[63021]: DEBUG nova.scheduler.client.report [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Updating ProviderTree inventory for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1110.765252] env[63021]: DEBUG nova.compute.provider_tree [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Updating inventory in ProviderTree for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1110.776465] env[63021]: DEBUG nova.scheduler.client.report [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Refreshing aggregate associations for resource provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d, aggregates: None {{(pid=63021) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1110.793946] env[63021]: DEBUG nova.scheduler.client.report [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Refreshing trait associations for resource provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE {{(pid=63021) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1110.857828] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34f4976f-c508-4f61-b05a-57142c3b5ad6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.865349] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-339a9c78-a0df-4c28-92f3-00090e97a5ab {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.893590] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb3a9d93-dc5e-445b-9bad-b80ae02448e6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.900845] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67dc8c3b-cbbd-4a93-a40a-a10f14d87c25 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.914573] env[63021]: DEBUG nova.compute.provider_tree [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Updating inventory in ProviderTree for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1111.244196] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 2209d709-7a5b-42e2-be93-b4fd436b6290] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1111.445924] env[63021]: DEBUG nova.scheduler.client.report [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Updated inventory for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with generation 133 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1111.446217] env[63021]: DEBUG nova.compute.provider_tree [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Updating resource provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d generation from 133 to 134 during operation: update_inventory {{(pid=63021) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1111.446401] env[63021]: DEBUG nova.compute.provider_tree [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Updating inventory in ProviderTree for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1111.747692] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: df0b6974-7f78-4b73-8583-d09754fbacea] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1111.951659] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.373s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.972089] env[63021]: INFO nova.scheduler.client.report [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Deleted allocations for instance ae8a512d-5393-432c-bca0-4dc86711891b [ 1112.251456] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 868b1aeb-8109-4cb9-bb58-33f3d92989ad] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1112.479345] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bdeab25c-70c0-40d3-91e8-322cb85fcad2 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "ae8a512d-5393-432c-bca0-4dc86711891b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.904s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1112.754460] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 36ec45ec-761b-4d62-a74f-e4d9a840ada0] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1113.258418] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: bef5c7cb-be2f-4be7-b72c-4a066d473dd0] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1113.529949] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "04ac365d-1f85-4903-8e5a-d4acc192d476" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.530209] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "04ac365d-1f85-4903-8e5a-d4acc192d476" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.761920] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 7c6d7839-616d-41f5-a909-f8c626477a5e] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1114.032137] env[63021]: DEBUG nova.compute.manager [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1114.265317] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 929cf4cd-6f3c-4988-8f80-d889d3070e1e] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1114.554197] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.554464] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.556457] env[63021]: INFO nova.compute.claims [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1114.772068] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 9a6720ef-5307-44b7-b566-8d553c9b5384] Instance has had 0 of 5 cleanup attempts {{(pid=63021) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1115.275482] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1115.275653] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Cleaning up deleted instances with incomplete migration {{(pid=63021) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1115.629887] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cb0933f-6b74-4495-9045-80232f000ef2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.638050] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5510ef94-e7bd-4ff0-98fc-1e717904843b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.667465] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d86fdf9-ad49-44de-a717-76d292a3eb2e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.674449] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-229ab2c6-dfbb-4d35-99a5-bac6a3043acb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.687977] env[63021]: DEBUG nova.compute.provider_tree [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1115.777610] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1116.191370] env[63021]: DEBUG nova.scheduler.client.report [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1116.696162] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.142s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.696700] env[63021]: DEBUG nova.compute.manager [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1117.202011] env[63021]: DEBUG nova.compute.utils [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1117.203496] env[63021]: DEBUG nova.compute.manager [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1117.203668] env[63021]: DEBUG nova.network.neutron [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1117.250515] env[63021]: DEBUG nova.policy [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ba16721213dd49a3b4e4990d7d319d1f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3c8959d2f1fe4706b7407c11e128b773', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 1117.279259] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1117.486103] env[63021]: DEBUG nova.network.neutron [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Successfully created port: e36495c3-444b-4ea4-aca2-dff752685aee {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1117.706620] env[63021]: DEBUG nova.compute.manager [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1117.784714] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1117.784714] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Starting heal instance info cache {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1118.716447] env[63021]: DEBUG nova.compute.manager [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1118.741599] env[63021]: DEBUG nova.virt.hardware [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1118.741861] env[63021]: DEBUG nova.virt.hardware [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1118.742033] env[63021]: DEBUG nova.virt.hardware [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1118.742220] env[63021]: DEBUG nova.virt.hardware [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1118.742371] env[63021]: DEBUG nova.virt.hardware [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1118.742519] env[63021]: DEBUG nova.virt.hardware [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1118.742728] env[63021]: DEBUG nova.virt.hardware [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1118.742887] env[63021]: DEBUG nova.virt.hardware [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1118.743064] env[63021]: DEBUG nova.virt.hardware [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1118.743229] env[63021]: DEBUG nova.virt.hardware [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1118.743403] env[63021]: DEBUG nova.virt.hardware [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1118.744346] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0592ad8e-1c23-465d-80b0-2c3c262c5a15 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.752338] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcf478d4-5164-4a61-b299-23963e524f2e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.846629] env[63021]: DEBUG nova.compute.manager [req-8cc6d27c-a3ed-4d88-adf9-06dbbdc239c6 req-91396551-83d4-4872-82b3-71af0e2d080a service nova] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Received event network-vif-plugged-e36495c3-444b-4ea4-aca2-dff752685aee {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1118.846866] env[63021]: DEBUG oslo_concurrency.lockutils [req-8cc6d27c-a3ed-4d88-adf9-06dbbdc239c6 req-91396551-83d4-4872-82b3-71af0e2d080a service nova] Acquiring lock "04ac365d-1f85-4903-8e5a-d4acc192d476-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.847094] env[63021]: DEBUG oslo_concurrency.lockutils [req-8cc6d27c-a3ed-4d88-adf9-06dbbdc239c6 req-91396551-83d4-4872-82b3-71af0e2d080a service nova] Lock "04ac365d-1f85-4903-8e5a-d4acc192d476-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.847264] env[63021]: DEBUG oslo_concurrency.lockutils [req-8cc6d27c-a3ed-4d88-adf9-06dbbdc239c6 req-91396551-83d4-4872-82b3-71af0e2d080a service nova] Lock "04ac365d-1f85-4903-8e5a-d4acc192d476-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.847507] env[63021]: DEBUG nova.compute.manager [req-8cc6d27c-a3ed-4d88-adf9-06dbbdc239c6 req-91396551-83d4-4872-82b3-71af0e2d080a service nova] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] No waiting events found dispatching network-vif-plugged-e36495c3-444b-4ea4-aca2-dff752685aee {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1118.847598] env[63021]: WARNING nova.compute.manager [req-8cc6d27c-a3ed-4d88-adf9-06dbbdc239c6 req-91396551-83d4-4872-82b3-71af0e2d080a service nova] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Received unexpected event network-vif-plugged-e36495c3-444b-4ea4-aca2-dff752685aee for instance with vm_state building and task_state spawning. [ 1118.928247] env[63021]: DEBUG nova.network.neutron [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Successfully updated port: e36495c3-444b-4ea4-aca2-dff752685aee {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1119.431083] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "refresh_cache-04ac365d-1f85-4903-8e5a-d4acc192d476" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1119.431265] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquired lock "refresh_cache-04ac365d-1f85-4903-8e5a-d4acc192d476" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1119.431393] env[63021]: DEBUG nova.network.neutron [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1119.690118] env[63021]: DEBUG nova.compute.manager [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Stashing vm_state: active {{(pid=63021) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1119.964105] env[63021]: DEBUG nova.network.neutron [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1120.087662] env[63021]: DEBUG nova.network.neutron [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Updating instance_info_cache with network_info: [{"id": "e36495c3-444b-4ea4-aca2-dff752685aee", "address": "fa:16:3e:ed:48:13", "network": {"id": "0c0dd1a7-6687-45c5-82b1-33f34bdbb848", "bridge": "br-int", "label": "tempest-ServersTestJSON-523240467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c8959d2f1fe4706b7407c11e128b773", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffc811e4-0e4e-4748-8c7e-b3f14ccbd42d", "external-id": "nsx-vlan-transportzone-404", "segmentation_id": 404, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape36495c3-44", "ovs_interfaceid": "e36495c3-444b-4ea4-aca2-dff752685aee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1120.209603] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.209901] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1120.296333] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Didn't find any instances for network info cache update. {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1120.296532] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1120.296681] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1120.296827] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1120.296981] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1120.590496] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Releasing lock "refresh_cache-04ac365d-1f85-4903-8e5a-d4acc192d476" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1120.590496] env[63021]: DEBUG nova.compute.manager [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Instance network_info: |[{"id": "e36495c3-444b-4ea4-aca2-dff752685aee", "address": "fa:16:3e:ed:48:13", "network": {"id": "0c0dd1a7-6687-45c5-82b1-33f34bdbb848", "bridge": "br-int", "label": "tempest-ServersTestJSON-523240467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c8959d2f1fe4706b7407c11e128b773", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffc811e4-0e4e-4748-8c7e-b3f14ccbd42d", "external-id": "nsx-vlan-transportzone-404", "segmentation_id": 404, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape36495c3-44", "ovs_interfaceid": "e36495c3-444b-4ea4-aca2-dff752685aee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1120.590889] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ed:48:13', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ffc811e4-0e4e-4748-8c7e-b3f14ccbd42d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e36495c3-444b-4ea4-aca2-dff752685aee', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1120.598336] env[63021]: DEBUG oslo.service.loopingcall [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1120.598656] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1120.598903] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-98a49313-31ce-4e21-a51b-c0af4a830f6a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.617979] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1120.617979] env[63021]: value = "task-1294352" [ 1120.617979] env[63021]: _type = "Task" [ 1120.617979] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.629277] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294352, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.715217] env[63021]: INFO nova.compute.claims [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1120.800349] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.872907] env[63021]: DEBUG nova.compute.manager [req-37d84afa-8ecb-4d85-9575-eeaae1104bcf req-a75f53e4-d1a3-45ec-a383-fa349762950d service nova] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Received event network-changed-e36495c3-444b-4ea4-aca2-dff752685aee {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1120.872907] env[63021]: DEBUG nova.compute.manager [req-37d84afa-8ecb-4d85-9575-eeaae1104bcf req-a75f53e4-d1a3-45ec-a383-fa349762950d service nova] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Refreshing instance network info cache due to event network-changed-e36495c3-444b-4ea4-aca2-dff752685aee. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1120.873115] env[63021]: DEBUG oslo_concurrency.lockutils [req-37d84afa-8ecb-4d85-9575-eeaae1104bcf req-a75f53e4-d1a3-45ec-a383-fa349762950d service nova] Acquiring lock "refresh_cache-04ac365d-1f85-4903-8e5a-d4acc192d476" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1120.873260] env[63021]: DEBUG oslo_concurrency.lockutils [req-37d84afa-8ecb-4d85-9575-eeaae1104bcf req-a75f53e4-d1a3-45ec-a383-fa349762950d service nova] Acquired lock "refresh_cache-04ac365d-1f85-4903-8e5a-d4acc192d476" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1120.873456] env[63021]: DEBUG nova.network.neutron [req-37d84afa-8ecb-4d85-9575-eeaae1104bcf req-a75f53e4-d1a3-45ec-a383-fa349762950d service nova] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Refreshing network info cache for port e36495c3-444b-4ea4-aca2-dff752685aee {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1121.128399] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294352, 'name': CreateVM_Task, 'duration_secs': 0.30232} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.128809] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1121.129204] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1121.129374] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1121.129686] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1121.129932] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e3dcba15-dd8f-4b45-a398-496c6e6a6056 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.134034] env[63021]: DEBUG oslo_vmware.api [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1121.134034] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52ef37fe-5944-f43d-c3e3-43f4f43f5ead" [ 1121.134034] env[63021]: _type = "Task" [ 1121.134034] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.141234] env[63021]: DEBUG oslo_vmware.api [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52ef37fe-5944-f43d-c3e3-43f4f43f5ead, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.221440] env[63021]: INFO nova.compute.resource_tracker [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Updating resource usage from migration ff919176-bc4f-4549-8596-4a1f9383b098 [ 1121.296265] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89138e86-8207-4274-b660-9737aecdff30 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.303931] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-767c0e40-da86-4855-b93d-27c2e780a302 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.335890] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08ea9e3b-a217-4fa9-9a31-284df4fb1436 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.342787] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de530772-e9f1-4cc7-9ea8-ab0e90d62250 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.355085] env[63021]: DEBUG nova.compute.provider_tree [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1121.554050] env[63021]: DEBUG nova.network.neutron [req-37d84afa-8ecb-4d85-9575-eeaae1104bcf req-a75f53e4-d1a3-45ec-a383-fa349762950d service nova] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Updated VIF entry in instance network info cache for port e36495c3-444b-4ea4-aca2-dff752685aee. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1121.554503] env[63021]: DEBUG nova.network.neutron [req-37d84afa-8ecb-4d85-9575-eeaae1104bcf req-a75f53e4-d1a3-45ec-a383-fa349762950d service nova] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Updating instance_info_cache with network_info: [{"id": "e36495c3-444b-4ea4-aca2-dff752685aee", "address": "fa:16:3e:ed:48:13", "network": {"id": "0c0dd1a7-6687-45c5-82b1-33f34bdbb848", "bridge": "br-int", "label": "tempest-ServersTestJSON-523240467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c8959d2f1fe4706b7407c11e128b773", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffc811e4-0e4e-4748-8c7e-b3f14ccbd42d", "external-id": "nsx-vlan-transportzone-404", "segmentation_id": 404, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape36495c3-44", "ovs_interfaceid": "e36495c3-444b-4ea4-aca2-dff752685aee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1121.644566] env[63021]: DEBUG oslo_vmware.api [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52ef37fe-5944-f43d-c3e3-43f4f43f5ead, 'name': SearchDatastore_Task, 'duration_secs': 0.010053} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.644875] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1121.645121] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1121.645369] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1121.645531] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1121.645715] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1121.645968] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e7f4fb49-1cbd-4353-b17c-60c3cd7e26b7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.653662] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1121.653806] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1121.654470] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c36efe83-cb45-48e2-8ced-ce56ddf3ca2f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.658938] env[63021]: DEBUG oslo_vmware.api [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1121.658938] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52402bb1-b4a1-6176-b43c-d78727d147f5" [ 1121.658938] env[63021]: _type = "Task" [ 1121.658938] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.667238] env[63021]: DEBUG oslo_vmware.api [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52402bb1-b4a1-6176-b43c-d78727d147f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.857805] env[63021]: DEBUG nova.scheduler.client.report [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1122.056711] env[63021]: DEBUG oslo_concurrency.lockutils [req-37d84afa-8ecb-4d85-9575-eeaae1104bcf req-a75f53e4-d1a3-45ec-a383-fa349762950d service nova] Releasing lock "refresh_cache-04ac365d-1f85-4903-8e5a-d4acc192d476" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1122.169200] env[63021]: DEBUG oslo_vmware.api [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52402bb1-b4a1-6176-b43c-d78727d147f5, 'name': SearchDatastore_Task, 'duration_secs': 0.008448} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.169934] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c95e4bda-ab17-451b-b6f9-0b0e438ae709 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.174870] env[63021]: DEBUG oslo_vmware.api [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1122.174870] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52e3ba96-9236-d28e-c44b-635bdc7c878f" [ 1122.174870] env[63021]: _type = "Task" [ 1122.174870] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.182095] env[63021]: DEBUG oslo_vmware.api [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52e3ba96-9236-d28e-c44b-635bdc7c878f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.362936] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.153s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.363178] env[63021]: INFO nova.compute.manager [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Migrating [ 1122.369429] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.569s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1122.369607] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.369754] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63021) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1122.372940] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce65df0e-fa59-42fc-bf4f-0f219e63878b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.385502] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22ca0b3e-29a1-464e-a59d-d1c031457931 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.400718] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a354059-0d52-433c-828f-155d739dcad2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.408240] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-276340c9-9372-4ecc-bb96-e973a4922e68 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.437080] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180041MB free_disk=139GB free_vcpus=48 pci_devices=None {{(pid=63021) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1122.437185] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.437369] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1122.685239] env[63021]: DEBUG oslo_vmware.api [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52e3ba96-9236-d28e-c44b-635bdc7c878f, 'name': SearchDatastore_Task, 'duration_secs': 0.008916} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.685500] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1122.685760] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 04ac365d-1f85-4903-8e5a-d4acc192d476/04ac365d-1f85-4903-8e5a-d4acc192d476.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1122.686018] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a8cb6ab0-06d0-4e5f-ba1c-6e2c78bbe93e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.692223] env[63021]: DEBUG oslo_vmware.api [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1122.692223] env[63021]: value = "task-1294353" [ 1122.692223] env[63021]: _type = "Task" [ 1122.692223] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.699513] env[63021]: DEBUG oslo_vmware.api [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294353, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.881279] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "refresh_cache-4041dfbf-fa94-4ff8-9176-e7088cdf241f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1122.881485] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquired lock "refresh_cache-4041dfbf-fa94-4ff8-9176-e7088cdf241f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1122.881672] env[63021]: DEBUG nova.network.neutron [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1123.203122] env[63021]: DEBUG oslo_vmware.api [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294353, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.462229} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.203122] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 04ac365d-1f85-4903-8e5a-d4acc192d476/04ac365d-1f85-4903-8e5a-d4acc192d476.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1123.203633] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1123.203633] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-becfca09-deda-402a-9186-36053859623f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.210031] env[63021]: DEBUG oslo_vmware.api [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1123.210031] env[63021]: value = "task-1294354" [ 1123.210031] env[63021]: _type = "Task" [ 1123.210031] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.218111] env[63021]: DEBUG oslo_vmware.api [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294354, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.444644] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Applying migration context for instance 4041dfbf-fa94-4ff8-9176-e7088cdf241f as it has an incoming, in-progress migration ff919176-bc4f-4549-8596-4a1f9383b098. Migration status is pre-migrating {{(pid=63021) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1123.445551] env[63021]: INFO nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Updating resource usage from migration ff919176-bc4f-4549-8596-4a1f9383b098 [ 1123.460142] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 129d8c8a-b869-422d-8e7f-cc3a9400021a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1123.460288] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance d03c47da-7e73-4474-ad9b-07d27629dcae actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1123.460409] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 02c217fb-174a-47f8-a9b7-d71497797590 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1123.460525] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 04ac365d-1f85-4903-8e5a-d4acc192d476 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1123.460643] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Migration ff919176-bc4f-4549-8596-4a1f9383b098 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1123.461009] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 4041dfbf-fa94-4ff8-9176-e7088cdf241f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1123.461009] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=63021) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1123.461116] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1728MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=63021) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1123.533277] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80cb8d90-9de0-4550-a96b-9aaca5483a0d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.540660] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a8c2120-c65e-499b-a2fa-ee7ef6b126f2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.570345] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-581e4e6e-0c5c-4f03-a304-0d40eebe7b12 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.576891] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fd89293-7ddc-4c43-be45-8d40961b488e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.590181] env[63021]: DEBUG nova.compute.provider_tree [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1123.602314] env[63021]: DEBUG nova.network.neutron [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Updating instance_info_cache with network_info: [{"id": "b9d1b591-e9bc-4e70-9c76-1a0673bca1d5", "address": "fa:16:3e:1b:b0:e6", "network": {"id": "9502e364-60e1-47ef-a1f7-f3e3ff45cf69", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-164717784-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5af87713ae6b477ca8a584aed287889d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9d1b591-e9", "ovs_interfaceid": "b9d1b591-e9bc-4e70-9c76-1a0673bca1d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1123.719161] env[63021]: DEBUG oslo_vmware.api [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294354, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065255} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.719436] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1123.720179] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cfdcf69-8b0e-423d-b1f9-643f580d1590 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.741440] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] 04ac365d-1f85-4903-8e5a-d4acc192d476/04ac365d-1f85-4903-8e5a-d4acc192d476.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1123.742033] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d3795fa2-e633-4831-9f18-2ba1be2aa47c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.760696] env[63021]: DEBUG oslo_vmware.api [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1123.760696] env[63021]: value = "task-1294355" [ 1123.760696] env[63021]: _type = "Task" [ 1123.760696] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.768136] env[63021]: DEBUG oslo_vmware.api [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294355, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.093225] env[63021]: DEBUG nova.scheduler.client.report [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1124.104240] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Releasing lock "refresh_cache-4041dfbf-fa94-4ff8-9176-e7088cdf241f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1124.270605] env[63021]: DEBUG oslo_vmware.api [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294355, 'name': ReconfigVM_Task, 'duration_secs': 0.331446} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.270983] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Reconfigured VM instance instance-0000006c to attach disk [datastore1] 04ac365d-1f85-4903-8e5a-d4acc192d476/04ac365d-1f85-4903-8e5a-d4acc192d476.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1124.271530] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7e3c7e79-24ab-4bab-942f-b6d67ae81c13 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.277137] env[63021]: DEBUG oslo_vmware.api [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1124.277137] env[63021]: value = "task-1294356" [ 1124.277137] env[63021]: _type = "Task" [ 1124.277137] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.284588] env[63021]: DEBUG oslo_vmware.api [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294356, 'name': Rename_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.598034] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63021) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1124.598239] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.161s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.786912] env[63021]: DEBUG oslo_vmware.api [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294356, 'name': Rename_Task, 'duration_secs': 0.149156} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.787218] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1124.787481] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3b96aae1-edbb-4f8f-9bc3-a41627126c38 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.793100] env[63021]: DEBUG oslo_vmware.api [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1124.793100] env[63021]: value = "task-1294357" [ 1124.793100] env[63021]: _type = "Task" [ 1124.793100] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.800937] env[63021]: DEBUG oslo_vmware.api [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294357, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.303061] env[63021]: DEBUG oslo_vmware.api [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294357, 'name': PowerOnVM_Task, 'duration_secs': 0.503611} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.303413] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1125.303496] env[63021]: INFO nova.compute.manager [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Took 6.59 seconds to spawn the instance on the hypervisor. [ 1125.303695] env[63021]: DEBUG nova.compute.manager [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1125.304463] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-268a8fbd-6566-4b75-a2b9-5f833649707b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.617793] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e4049f8-25f1-45b4-b7a2-4d0f71a9241b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.638322] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Updating instance '4041dfbf-fa94-4ff8-9176-e7088cdf241f' progress to 0 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1125.821816] env[63021]: INFO nova.compute.manager [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Took 11.28 seconds to build instance. [ 1126.144259] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1126.144586] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eae9df85-07e3-4816-94db-34de277b3bf5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.152687] env[63021]: DEBUG oslo_vmware.api [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1126.152687] env[63021]: value = "task-1294358" [ 1126.152687] env[63021]: _type = "Task" [ 1126.152687] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.160159] env[63021]: DEBUG oslo_vmware.api [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294358, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.323329] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f908d173-a792-4f72-9fa1-07f5a2540d2a tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "04ac365d-1f85-4903-8e5a-d4acc192d476" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.793s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.662678] env[63021]: DEBUG oslo_vmware.api [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294358, 'name': PowerOffVM_Task, 'duration_secs': 0.190569} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.662986] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1126.663198] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Updating instance '4041dfbf-fa94-4ff8-9176-e7088cdf241f' progress to 17 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1126.807171] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bdbda6fb-d052-4f65-8841-e1625dc578e6 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "04ac365d-1f85-4903-8e5a-d4acc192d476" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1126.807501] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bdbda6fb-d052-4f65-8841-e1625dc578e6 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "04ac365d-1f85-4903-8e5a-d4acc192d476" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1126.807707] env[63021]: DEBUG nova.compute.manager [None req-bdbda6fb-d052-4f65-8841-e1625dc578e6 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1126.808637] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e942e393-6b39-453b-aafd-71c305affb26 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.815528] env[63021]: DEBUG nova.compute.manager [None req-bdbda6fb-d052-4f65-8841-e1625dc578e6 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63021) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1126.816084] env[63021]: DEBUG nova.objects.instance [None req-bdbda6fb-d052-4f65-8841-e1625dc578e6 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lazy-loading 'flavor' on Instance uuid 04ac365d-1f85-4903-8e5a-d4acc192d476 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1127.169787] env[63021]: DEBUG nova.virt.hardware [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1127.170049] env[63021]: DEBUG nova.virt.hardware [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1127.170221] env[63021]: DEBUG nova.virt.hardware [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1127.170406] env[63021]: DEBUG nova.virt.hardware [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1127.170559] env[63021]: DEBUG nova.virt.hardware [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1127.170711] env[63021]: DEBUG nova.virt.hardware [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1127.170930] env[63021]: DEBUG nova.virt.hardware [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1127.171097] env[63021]: DEBUG nova.virt.hardware [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1127.171270] env[63021]: DEBUG nova.virt.hardware [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1127.171439] env[63021]: DEBUG nova.virt.hardware [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1127.171617] env[63021]: DEBUG nova.virt.hardware [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1127.176684] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-500e36bb-1612-4265-966e-2d456d97c30e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.193067] env[63021]: DEBUG oslo_vmware.api [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1127.193067] env[63021]: value = "task-1294359" [ 1127.193067] env[63021]: _type = "Task" [ 1127.193067] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.200745] env[63021]: DEBUG oslo_vmware.api [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294359, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.320965] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdbda6fb-d052-4f65-8841-e1625dc578e6 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1127.321249] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-629cdea2-4389-417f-ba96-1fc94153fcb0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.327329] env[63021]: DEBUG oslo_vmware.api [None req-bdbda6fb-d052-4f65-8841-e1625dc578e6 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1127.327329] env[63021]: value = "task-1294360" [ 1127.327329] env[63021]: _type = "Task" [ 1127.327329] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.335035] env[63021]: DEBUG oslo_vmware.api [None req-bdbda6fb-d052-4f65-8841-e1625dc578e6 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294360, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.704990] env[63021]: DEBUG oslo_vmware.api [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294359, 'name': ReconfigVM_Task, 'duration_secs': 0.186041} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.705387] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Updating instance '4041dfbf-fa94-4ff8-9176-e7088cdf241f' progress to 33 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1127.837617] env[63021]: DEBUG oslo_vmware.api [None req-bdbda6fb-d052-4f65-8841-e1625dc578e6 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294360, 'name': PowerOffVM_Task, 'duration_secs': 0.199751} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.837900] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdbda6fb-d052-4f65-8841-e1625dc578e6 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1127.838108] env[63021]: DEBUG nova.compute.manager [None req-bdbda6fb-d052-4f65-8841-e1625dc578e6 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1127.838956] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-882ccf2d-6a86-4747-a3f1-c9d7710e67fd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.212370] env[63021]: DEBUG nova.virt.hardware [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1128.212621] env[63021]: DEBUG nova.virt.hardware [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1128.212783] env[63021]: DEBUG nova.virt.hardware [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1128.212961] env[63021]: DEBUG nova.virt.hardware [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1128.213158] env[63021]: DEBUG nova.virt.hardware [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1128.213313] env[63021]: DEBUG nova.virt.hardware [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1128.213594] env[63021]: DEBUG nova.virt.hardware [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1128.213765] env[63021]: DEBUG nova.virt.hardware [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1128.213935] env[63021]: DEBUG nova.virt.hardware [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1128.214115] env[63021]: DEBUG nova.virt.hardware [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1128.214294] env[63021]: DEBUG nova.virt.hardware [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1128.219694] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Reconfiguring VM instance instance-00000069 to detach disk 2000 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1128.219981] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e1003f10-672e-4a89-bfe5-1e84e05234e1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.238484] env[63021]: DEBUG oslo_vmware.api [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1128.238484] env[63021]: value = "task-1294361" [ 1128.238484] env[63021]: _type = "Task" [ 1128.238484] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.246532] env[63021]: DEBUG oslo_vmware.api [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294361, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.351256] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bdbda6fb-d052-4f65-8841-e1625dc578e6 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "04ac365d-1f85-4903-8e5a-d4acc192d476" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.544s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.747903] env[63021]: DEBUG oslo_vmware.api [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294361, 'name': ReconfigVM_Task, 'duration_secs': 0.154117} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.748209] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Reconfigured VM instance instance-00000069 to detach disk 2000 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1128.749189] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebc01fe5-405b-4802-a078-e04b30994cbb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.770713] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] 4041dfbf-fa94-4ff8-9176-e7088cdf241f/4041dfbf-fa94-4ff8-9176-e7088cdf241f.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1128.770992] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9c5d5bda-2d6f-4f08-867e-16e3fe7cf79e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.789596] env[63021]: DEBUG oslo_vmware.api [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1128.789596] env[63021]: value = "task-1294362" [ 1128.789596] env[63021]: _type = "Task" [ 1128.789596] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.797015] env[63021]: DEBUG oslo_vmware.api [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294362, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.298845] env[63021]: DEBUG oslo_vmware.api [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294362, 'name': ReconfigVM_Task, 'duration_secs': 0.252556} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.299211] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Reconfigured VM instance instance-00000069 to attach disk [datastore2] 4041dfbf-fa94-4ff8-9176-e7088cdf241f/4041dfbf-fa94-4ff8-9176-e7088cdf241f.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1129.299528] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Updating instance '4041dfbf-fa94-4ff8-9176-e7088cdf241f' progress to 50 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1129.446708] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "04ac365d-1f85-4903-8e5a-d4acc192d476" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.447079] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "04ac365d-1f85-4903-8e5a-d4acc192d476" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.447155] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "04ac365d-1f85-4903-8e5a-d4acc192d476-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.447343] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "04ac365d-1f85-4903-8e5a-d4acc192d476-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.447577] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "04ac365d-1f85-4903-8e5a-d4acc192d476-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.450117] env[63021]: INFO nova.compute.manager [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Terminating instance [ 1129.451807] env[63021]: DEBUG nova.compute.manager [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1129.452013] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1129.452836] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c237f46-8402-49a9-96ea-55149e840037 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.460205] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1129.460685] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b099c6dc-2856-4c4c-8ff8-d49984cdd3c1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.525726] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1129.525955] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1129.526156] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Deleting the datastore file [datastore1] 04ac365d-1f85-4903-8e5a-d4acc192d476 {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1129.526425] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0af157ef-e217-48e3-af44-bf27f92ea80e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.533408] env[63021]: DEBUG oslo_vmware.api [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1129.533408] env[63021]: value = "task-1294364" [ 1129.533408] env[63021]: _type = "Task" [ 1129.533408] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.540682] env[63021]: DEBUG oslo_vmware.api [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294364, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.806099] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c1ea92c-a294-4d7e-a37d-157014afedff {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.824757] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ee22077-60a5-483f-acdc-aa4a115cd346 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.840950] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Updating instance '4041dfbf-fa94-4ff8-9176-e7088cdf241f' progress to 67 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1130.042922] env[63021]: DEBUG oslo_vmware.api [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294364, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157754} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.043214] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1130.043401] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1130.043576] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1130.043750] env[63021]: INFO nova.compute.manager [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Took 0.59 seconds to destroy the instance on the hypervisor. [ 1130.043982] env[63021]: DEBUG oslo.service.loopingcall [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1130.044196] env[63021]: DEBUG nova.compute.manager [-] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1130.044290] env[63021]: DEBUG nova.network.neutron [-] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1130.284170] env[63021]: DEBUG nova.compute.manager [req-8084ffb4-9052-4705-b48a-7c29aa757481 req-5bcf10fe-d5e0-4a6b-9011-bd0ae0f0d4b2 service nova] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Received event network-vif-deleted-e36495c3-444b-4ea4-aca2-dff752685aee {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1130.284368] env[63021]: INFO nova.compute.manager [req-8084ffb4-9052-4705-b48a-7c29aa757481 req-5bcf10fe-d5e0-4a6b-9011-bd0ae0f0d4b2 service nova] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Neutron deleted interface e36495c3-444b-4ea4-aca2-dff752685aee; detaching it from the instance and deleting it from the info cache [ 1130.284543] env[63021]: DEBUG nova.network.neutron [req-8084ffb4-9052-4705-b48a-7c29aa757481 req-5bcf10fe-d5e0-4a6b-9011-bd0ae0f0d4b2 service nova] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1130.379422] env[63021]: DEBUG nova.network.neutron [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Port b9d1b591-e9bc-4e70-9c76-1a0673bca1d5 binding to destination host cpu-1 is already ACTIVE {{(pid=63021) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1130.762714] env[63021]: DEBUG nova.network.neutron [-] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1130.786948] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ee930d58-329f-43b5-946f-440a88c9fb0d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.796157] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c954dee-bef0-489e-a584-cac6deed78d0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.822459] env[63021]: DEBUG nova.compute.manager [req-8084ffb4-9052-4705-b48a-7c29aa757481 req-5bcf10fe-d5e0-4a6b-9011-bd0ae0f0d4b2 service nova] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Detach interface failed, port_id=e36495c3-444b-4ea4-aca2-dff752685aee, reason: Instance 04ac365d-1f85-4903-8e5a-d4acc192d476 could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1131.265315] env[63021]: INFO nova.compute.manager [-] [instance: 04ac365d-1f85-4903-8e5a-d4acc192d476] Took 1.22 seconds to deallocate network for instance. [ 1131.401390] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "4041dfbf-fa94-4ff8-9176-e7088cdf241f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1131.401614] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "4041dfbf-fa94-4ff8-9176-e7088cdf241f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.401743] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "4041dfbf-fa94-4ff8-9176-e7088cdf241f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.772196] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1131.772542] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.772710] env[63021]: DEBUG nova.objects.instance [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lazy-loading 'resources' on Instance uuid 04ac365d-1f85-4903-8e5a-d4acc192d476 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1132.351743] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13f1a0cd-86cb-4a4f-b31f-fd5f580d8738 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.359105] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5569c5d-1cac-4ebd-8090-947718f392d4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.389106] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7faa936f-faf3-460e-8394-d3ec944169c4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.395699] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3dd00bd-8f55-4c58-b96b-a1faa96b81c7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.411750] env[63021]: DEBUG nova.compute.provider_tree [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Updating inventory in ProviderTree for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1132.440708] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "refresh_cache-4041dfbf-fa94-4ff8-9176-e7088cdf241f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1132.440900] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquired lock "refresh_cache-4041dfbf-fa94-4ff8-9176-e7088cdf241f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1132.441087] env[63021]: DEBUG nova.network.neutron [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1132.933079] env[63021]: ERROR nova.scheduler.client.report [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [req-fa743e2a-2535-4557-9805-3cfcfedbece4] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 6ef6c041-2b0d-4bc3-8662-5d679a77a86d. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-fa743e2a-2535-4557-9805-3cfcfedbece4"}]} [ 1132.951803] env[63021]: DEBUG nova.scheduler.client.report [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Refreshing inventories for resource provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1132.965943] env[63021]: DEBUG nova.scheduler.client.report [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Updating ProviderTree inventory for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1132.966208] env[63021]: DEBUG nova.compute.provider_tree [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Updating inventory in ProviderTree for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1132.977850] env[63021]: DEBUG nova.scheduler.client.report [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Refreshing aggregate associations for resource provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d, aggregates: None {{(pid=63021) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1132.996800] env[63021]: DEBUG nova.scheduler.client.report [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Refreshing trait associations for resource provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE {{(pid=63021) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1133.072232] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b17aeb68-e0cd-4d60-83c5-82886d476a70 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.080117] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ab89a3e-cc0f-478d-bb2d-48fa58b8a96b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.111697] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdb73440-82ac-41a9-9bc8-8a61d59f4d3f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.119116] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1094368-2ca0-41d9-b01e-82e1a43927ad {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.131986] env[63021]: DEBUG nova.compute.provider_tree [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Updating inventory in ProviderTree for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1133.178300] env[63021]: DEBUG nova.network.neutron [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Updating instance_info_cache with network_info: [{"id": "b9d1b591-e9bc-4e70-9c76-1a0673bca1d5", "address": "fa:16:3e:1b:b0:e6", "network": {"id": "9502e364-60e1-47ef-a1f7-f3e3ff45cf69", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-164717784-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5af87713ae6b477ca8a584aed287889d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9d1b591-e9", "ovs_interfaceid": "b9d1b591-e9bc-4e70-9c76-1a0673bca1d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1133.660771] env[63021]: DEBUG nova.scheduler.client.report [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Updated inventory for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with generation 137 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1133.661072] env[63021]: DEBUG nova.compute.provider_tree [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Updating resource provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d generation from 137 to 138 during operation: update_inventory {{(pid=63021) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1133.661266] env[63021]: DEBUG nova.compute.provider_tree [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Updating inventory in ProviderTree for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1133.680668] env[63021]: DEBUG oslo_concurrency.lockutils [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Releasing lock "refresh_cache-4041dfbf-fa94-4ff8-9176-e7088cdf241f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1134.166570] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.394s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1134.189785] env[63021]: INFO nova.scheduler.client.report [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Deleted allocations for instance 04ac365d-1f85-4903-8e5a-d4acc192d476 [ 1134.211069] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9934f8fe-a18f-4454-b194-13e3631e4e71 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.237972] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fecbe8c-de5c-4f35-b5ab-f249b15f638f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.248113] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Updating instance '4041dfbf-fa94-4ff8-9176-e7088cdf241f' progress to 83 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1134.698697] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bc5bfea0-cbc2-4df7-b47e-653e7c3f8db4 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "04ac365d-1f85-4903-8e5a-d4acc192d476" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.252s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1134.754164] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1134.754164] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2d04a1cb-7cec-4970-9616-e2eef408068c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.762052] env[63021]: DEBUG oslo_vmware.api [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1134.762052] env[63021]: value = "task-1294365" [ 1134.762052] env[63021]: _type = "Task" [ 1134.762052] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.772235] env[63021]: DEBUG oslo_vmware.api [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294365, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.780144] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "02c217fb-174a-47f8-a9b7-d71497797590" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1134.780394] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "02c217fb-174a-47f8-a9b7-d71497797590" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1134.780602] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "02c217fb-174a-47f8-a9b7-d71497797590-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1134.780789] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "02c217fb-174a-47f8-a9b7-d71497797590-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1134.780957] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "02c217fb-174a-47f8-a9b7-d71497797590-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1134.783070] env[63021]: INFO nova.compute.manager [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Terminating instance [ 1134.784849] env[63021]: DEBUG nova.compute.manager [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1134.785079] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1134.785874] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77f2fb9a-c02e-46a3-82dd-58513a87a46a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.793138] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1134.793420] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1f8b6d36-43a8-42ca-8379-96de73bfda69 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.800292] env[63021]: DEBUG oslo_vmware.api [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1134.800292] env[63021]: value = "task-1294366" [ 1134.800292] env[63021]: _type = "Task" [ 1134.800292] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.808858] env[63021]: DEBUG oslo_vmware.api [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294366, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.172939] env[63021]: DEBUG oslo_concurrency.lockutils [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "129d8c8a-b869-422d-8e7f-cc3a9400021a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1135.173399] env[63021]: DEBUG oslo_concurrency.lockutils [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "129d8c8a-b869-422d-8e7f-cc3a9400021a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1135.173447] env[63021]: DEBUG oslo_concurrency.lockutils [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "129d8c8a-b869-422d-8e7f-cc3a9400021a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1135.173636] env[63021]: DEBUG oslo_concurrency.lockutils [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "129d8c8a-b869-422d-8e7f-cc3a9400021a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1135.173814] env[63021]: DEBUG oslo_concurrency.lockutils [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "129d8c8a-b869-422d-8e7f-cc3a9400021a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1135.176026] env[63021]: INFO nova.compute.manager [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Terminating instance [ 1135.177784] env[63021]: DEBUG nova.compute.manager [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1135.177975] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1135.178796] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c62049c8-3647-48cd-b8b5-2ddd024c610b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.187655] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1135.187877] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8fdfe89a-df98-4bb0-84bd-938384a00c9f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.193598] env[63021]: DEBUG oslo_vmware.api [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1135.193598] env[63021]: value = "task-1294367" [ 1135.193598] env[63021]: _type = "Task" [ 1135.193598] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.201221] env[63021]: DEBUG oslo_vmware.api [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294367, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.274585] env[63021]: DEBUG oslo_vmware.api [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294365, 'name': PowerOnVM_Task, 'duration_secs': 0.362685} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.274983] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1135.275271] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-f9e2c837-ab60-4b09-b69b-187aa81fc534 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Updating instance '4041dfbf-fa94-4ff8-9176-e7088cdf241f' progress to 100 {{(pid=63021) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1135.311657] env[63021]: DEBUG oslo_vmware.api [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294366, 'name': PowerOffVM_Task, 'duration_secs': 0.160325} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.311996] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1135.312248] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1135.312543] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b5dcbbf8-e4b0-4679-8cc5-f1c7555acd97 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.386662] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1135.387607] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1135.387607] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Deleting the datastore file [datastore1] 02c217fb-174a-47f8-a9b7-d71497797590 {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1135.387607] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d1cb55b4-ee3d-47e6-8296-0aa58b5353b7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.394192] env[63021]: DEBUG oslo_vmware.api [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1135.394192] env[63021]: value = "task-1294369" [ 1135.394192] env[63021]: _type = "Task" [ 1135.394192] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.401640] env[63021]: DEBUG oslo_vmware.api [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294369, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.703969] env[63021]: DEBUG oslo_vmware.api [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294367, 'name': PowerOffVM_Task, 'duration_secs': 0.214511} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.704230] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1135.704403] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1135.704697] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0c24504a-9970-426d-9172-0d6c307ea56f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.784126] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1135.784354] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Deleting contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1135.784535] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Deleting the datastore file [datastore2] 129d8c8a-b869-422d-8e7f-cc3a9400021a {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1135.784998] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3c719d1b-34c6-469c-98f3-a25fed9d012d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.791334] env[63021]: DEBUG oslo_vmware.api [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for the task: (returnval){ [ 1135.791334] env[63021]: value = "task-1294371" [ 1135.791334] env[63021]: _type = "Task" [ 1135.791334] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.799936] env[63021]: DEBUG oslo_vmware.api [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294371, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.904352] env[63021]: DEBUG oslo_vmware.api [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294369, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.184126} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.904657] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1135.904885] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1135.905107] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1135.905305] env[63021]: INFO nova.compute.manager [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1135.905599] env[63021]: DEBUG oslo.service.loopingcall [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1135.905819] env[63021]: DEBUG nova.compute.manager [-] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1135.905929] env[63021]: DEBUG nova.network.neutron [-] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1136.301229] env[63021]: DEBUG oslo_vmware.api [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Task: {'id': task-1294371, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.323643} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.301582] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1136.301668] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Deleted contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1136.301841] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1136.302023] env[63021]: INFO nova.compute.manager [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1136.302256] env[63021]: DEBUG oslo.service.loopingcall [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1136.302449] env[63021]: DEBUG nova.compute.manager [-] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1136.302544] env[63021]: DEBUG nova.network.neutron [-] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1136.361497] env[63021]: DEBUG nova.compute.manager [req-f407e2ef-1c00-4efe-a75b-b72d3a5ec635 req-e624d901-a63b-4a33-a4b2-e17cc30ee384 service nova] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Received event network-vif-deleted-416fe99a-8969-4bb8-90d8-b9c9708917ca {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1136.361734] env[63021]: INFO nova.compute.manager [req-f407e2ef-1c00-4efe-a75b-b72d3a5ec635 req-e624d901-a63b-4a33-a4b2-e17cc30ee384 service nova] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Neutron deleted interface 416fe99a-8969-4bb8-90d8-b9c9708917ca; detaching it from the instance and deleting it from the info cache [ 1136.361923] env[63021]: DEBUG nova.network.neutron [req-f407e2ef-1c00-4efe-a75b-b72d3a5ec635 req-e624d901-a63b-4a33-a4b2-e17cc30ee384 service nova] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1136.840501] env[63021]: DEBUG nova.network.neutron [-] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1136.864534] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-19b42967-fdd1-4e8e-b0c6-3f2b0a931d65 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.874582] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44572e80-c933-4001-a375-6eab24e49096 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.897874] env[63021]: DEBUG nova.compute.manager [req-f407e2ef-1c00-4efe-a75b-b72d3a5ec635 req-e624d901-a63b-4a33-a4b2-e17cc30ee384 service nova] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Detach interface failed, port_id=416fe99a-8969-4bb8-90d8-b9c9708917ca, reason: Instance 02c217fb-174a-47f8-a9b7-d71497797590 could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1137.036720] env[63021]: DEBUG nova.network.neutron [-] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1137.342788] env[63021]: INFO nova.compute.manager [-] [instance: 02c217fb-174a-47f8-a9b7-d71497797590] Took 1.44 seconds to deallocate network for instance. [ 1137.345499] env[63021]: DEBUG nova.network.neutron [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Port b9d1b591-e9bc-4e70-9c76-1a0673bca1d5 binding to destination host cpu-1 is already ACTIVE {{(pid=63021) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1137.345736] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "refresh_cache-4041dfbf-fa94-4ff8-9176-e7088cdf241f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1137.345887] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquired lock "refresh_cache-4041dfbf-fa94-4ff8-9176-e7088cdf241f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1137.346059] env[63021]: DEBUG nova.network.neutron [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1137.539055] env[63021]: INFO nova.compute.manager [-] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Took 1.24 seconds to deallocate network for instance. [ 1137.853192] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.853462] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.853686] env[63021]: DEBUG nova.objects.instance [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lazy-loading 'resources' on Instance uuid 02c217fb-174a-47f8-a9b7-d71497797590 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1138.046861] env[63021]: DEBUG oslo_concurrency.lockutils [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.056418] env[63021]: DEBUG nova.network.neutron [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Updating instance_info_cache with network_info: [{"id": "b9d1b591-e9bc-4e70-9c76-1a0673bca1d5", "address": "fa:16:3e:1b:b0:e6", "network": {"id": "9502e364-60e1-47ef-a1f7-f3e3ff45cf69", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-164717784-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5af87713ae6b477ca8a584aed287889d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9d1b591-e9", "ovs_interfaceid": "b9d1b591-e9bc-4e70-9c76-1a0673bca1d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1138.393571] env[63021]: DEBUG nova.compute.manager [req-6bfb9c56-060d-48de-b13d-7b01bc8b2df7 req-4140a928-2d81-4dd2-a345-c6ef2fc362d0 service nova] [instance: 129d8c8a-b869-422d-8e7f-cc3a9400021a] Received event network-vif-deleted-99e60085-e84c-4670-9b34-3ceb87f8db12 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1138.426443] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4927bcf2-4057-4149-a8ab-8a5e22cfc122 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.433482] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1a679af-0394-4a17-b6a2-dfade627bc0b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.466612] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47e4527f-4ae2-43be-affd-01689689959b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.473576] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acb7312f-1b98-480d-9f99-628e26fb2a39 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.486195] env[63021]: DEBUG nova.compute.provider_tree [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Updating inventory in ProviderTree for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1138.559069] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Releasing lock "refresh_cache-4041dfbf-fa94-4ff8-9176-e7088cdf241f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1139.017559] env[63021]: DEBUG nova.scheduler.client.report [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Updated inventory for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with generation 138 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1139.017854] env[63021]: DEBUG nova.compute.provider_tree [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Updating resource provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d generation from 138 to 139 during operation: update_inventory {{(pid=63021) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1139.018049] env[63021]: DEBUG nova.compute.provider_tree [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Updating inventory in ProviderTree for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1139.062343] env[63021]: DEBUG nova.compute.manager [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=63021) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1139.062569] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1139.499024] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._sync_power_states {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1139.522512] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.669s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1139.524674] env[63021]: DEBUG oslo_concurrency.lockutils [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.478s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.524910] env[63021]: DEBUG nova.objects.instance [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lazy-loading 'resources' on Instance uuid 129d8c8a-b869-422d-8e7f-cc3a9400021a {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1139.541180] env[63021]: INFO nova.scheduler.client.report [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Deleted allocations for instance 02c217fb-174a-47f8-a9b7-d71497797590 [ 1140.003408] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Getting list of instances from cluster (obj){ [ 1140.003408] env[63021]: value = "domain-c8" [ 1140.003408] env[63021]: _type = "ClusterComputeResource" [ 1140.003408] env[63021]: } {{(pid=63021) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1140.004518] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b5ca8b7-80fc-4237-92c6-7d0c5477b4e5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.015259] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Got total of 2 instances {{(pid=63021) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1140.015516] env[63021]: WARNING nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] While synchronizing instance power states, found 4 instances in the database and 2 instances on the hypervisor. [ 1140.015694] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Triggering sync for uuid 129d8c8a-b869-422d-8e7f-cc3a9400021a {{(pid=63021) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1140.015891] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Triggering sync for uuid d03c47da-7e73-4474-ad9b-07d27629dcae {{(pid=63021) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1140.016062] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Triggering sync for uuid 4041dfbf-fa94-4ff8-9176-e7088cdf241f {{(pid=63021) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1140.016243] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Triggering sync for uuid 02c217fb-174a-47f8-a9b7-d71497797590 {{(pid=63021) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1140.016625] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquiring lock "129d8c8a-b869-422d-8e7f-cc3a9400021a" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.016911] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquiring lock "d03c47da-7e73-4474-ad9b-07d27629dcae" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.017140] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "d03c47da-7e73-4474-ad9b-07d27629dcae" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.017432] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquiring lock "4041dfbf-fa94-4ff8-9176-e7088cdf241f" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.017640] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "4041dfbf-fa94-4ff8-9176-e7088cdf241f" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.017814] env[63021]: INFO nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] During sync_power_state the instance has a pending task (resize_reverting). Skip. [ 1140.017979] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "4041dfbf-fa94-4ff8-9176-e7088cdf241f" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.018183] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquiring lock "02c217fb-174a-47f8-a9b7-d71497797590" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.019030] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00f4c6ed-28c3-4716-88fb-1981e1fe65dd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.048367] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5ba73280-e0ae-457b-88c6-fa76ee8898e0 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "02c217fb-174a-47f8-a9b7-d71497797590" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.268s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.049360] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "02c217fb-174a-47f8-a9b7-d71497797590" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.031s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.055282] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bd2a561a-b7c1-4c36-a9a8-63e1b8204807 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.067198] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9058cdbd-2261-4cf1-a6e4-d77d48836798 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.126215] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cadb116-ec2e-47f3-92af-27df281633be {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.133850] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d38ab5e-d0dd-4024-aab5-57726d51e17d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.163365] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18fbab7b-2746-4bf0-8356-5673fb0add9f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.171031] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eaa55d1-54f2-4ea8-95e8-c4ec524bc503 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.183886] env[63021]: DEBUG nova.compute.provider_tree [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1140.533896] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "d03c47da-7e73-4474-ad9b-07d27629dcae" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.516s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.604378] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "02c217fb-174a-47f8-a9b7-d71497797590" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.555s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.686683] env[63021]: DEBUG nova.scheduler.client.report [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1140.746948] env[63021]: DEBUG oslo_concurrency.lockutils [None req-58c14eeb-fc62-43f9-85d8-3e091c80fc50 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "d03c47da-7e73-4474-ad9b-07d27629dcae" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.747252] env[63021]: DEBUG oslo_concurrency.lockutils [None req-58c14eeb-fc62-43f9-85d8-3e091c80fc50 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "d03c47da-7e73-4474-ad9b-07d27629dcae" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.191454] env[63021]: DEBUG oslo_concurrency.lockutils [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.667s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1141.193801] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 2.131s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.211248] env[63021]: INFO nova.scheduler.client.report [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Deleted allocations for instance 129d8c8a-b869-422d-8e7f-cc3a9400021a [ 1141.249919] env[63021]: INFO nova.compute.manager [None req-58c14eeb-fc62-43f9-85d8-3e091c80fc50 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Detaching volume cd524289-5d08-4b7f-a083-259e501f1fc1 [ 1141.340473] env[63021]: INFO nova.virt.block_device [None req-58c14eeb-fc62-43f9-85d8-3e091c80fc50 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Attempting to driver detach volume cd524289-5d08-4b7f-a083-259e501f1fc1 from mountpoint /dev/sdb [ 1141.340730] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-58c14eeb-fc62-43f9-85d8-3e091c80fc50 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Volume detach. Driver type: vmdk {{(pid=63021) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1141.340931] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-58c14eeb-fc62-43f9-85d8-3e091c80fc50 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277598', 'volume_id': 'cd524289-5d08-4b7f-a083-259e501f1fc1', 'name': 'volume-cd524289-5d08-4b7f-a083-259e501f1fc1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd03c47da-7e73-4474-ad9b-07d27629dcae', 'attached_at': '', 'detached_at': '', 'volume_id': 'cd524289-5d08-4b7f-a083-259e501f1fc1', 'serial': 'cd524289-5d08-4b7f-a083-259e501f1fc1'} {{(pid=63021) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1141.341916] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f918e0b-6a5f-4f2b-b8d9-fac8e0b6f06b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.363770] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71fc3349-5989-4566-9574-34fdaee2e9f5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.370721] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7c099a5-8a58-4698-902c-b8c8df5c8b4d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.391015] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a004c5a-0d49-4358-a691-81fe3172f27b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.405476] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-58c14eeb-fc62-43f9-85d8-3e091c80fc50 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] The volume has not been displaced from its original location: [datastore2] volume-cd524289-5d08-4b7f-a083-259e501f1fc1/volume-cd524289-5d08-4b7f-a083-259e501f1fc1.vmdk. No consolidation needed. {{(pid=63021) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1141.410743] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-58c14eeb-fc62-43f9-85d8-3e091c80fc50 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Reconfiguring VM instance instance-00000062 to detach disk 2001 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1141.411073] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f7d24a5b-9957-44ba-9b61-3d36f28ebc79 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.435305] env[63021]: DEBUG oslo_vmware.api [None req-58c14eeb-fc62-43f9-85d8-3e091c80fc50 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1141.435305] env[63021]: value = "task-1294372" [ 1141.435305] env[63021]: _type = "Task" [ 1141.435305] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.444638] env[63021]: DEBUG oslo_vmware.api [None req-58c14eeb-fc62-43f9-85d8-3e091c80fc50 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294372, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.696654] env[63021]: DEBUG nova.objects.instance [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lazy-loading 'migration_context' on Instance uuid 4041dfbf-fa94-4ff8-9176-e7088cdf241f {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1141.717591] env[63021]: DEBUG oslo_concurrency.lockutils [None req-75532784-317e-4b6f-a770-98ddd49239f5 tempest-ServersTestJSON-2069293020 tempest-ServersTestJSON-2069293020-project-member] Lock "129d8c8a-b869-422d-8e7f-cc3a9400021a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.544s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1141.718407] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "129d8c8a-b869-422d-8e7f-cc3a9400021a" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 1.702s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.718854] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1c5d85c9-c545-45be-b82a-22dd2188988b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.728120] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-657c3026-5c39-4c12-897e-d4e009cdcf81 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.945786] env[63021]: DEBUG oslo_vmware.api [None req-58c14eeb-fc62-43f9-85d8-3e091c80fc50 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294372, 'name': ReconfigVM_Task, 'duration_secs': 0.255395} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.946087] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-58c14eeb-fc62-43f9-85d8-3e091c80fc50 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Reconfigured VM instance instance-00000062 to detach disk 2001 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1141.950666] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9f30cc2b-1af3-4f61-b967-943e66ce18a2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.964967] env[63021]: DEBUG oslo_vmware.api [None req-58c14eeb-fc62-43f9-85d8-3e091c80fc50 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1141.964967] env[63021]: value = "task-1294373" [ 1141.964967] env[63021]: _type = "Task" [ 1141.964967] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.972482] env[63021]: DEBUG oslo_vmware.api [None req-58c14eeb-fc62-43f9-85d8-3e091c80fc50 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294373, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.252033] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72b4b1f7-c3ab-4232-8ab1-9cf88a80e7f2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.255080] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "129d8c8a-b869-422d-8e7f-cc3a9400021a" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.537s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.260733] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64ce7051-8919-463a-9457-1d972292d142 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.290729] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57285dbb-6ce2-4cf4-9f79-9efeea0e1daa {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.297707] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a062ad94-bf23-41b6-a8ea-b27e6330d3da {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.312919] env[63021]: DEBUG nova.compute.provider_tree [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1142.474411] env[63021]: DEBUG oslo_vmware.api [None req-58c14eeb-fc62-43f9-85d8-3e091c80fc50 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294373, 'name': ReconfigVM_Task, 'duration_secs': 0.134705} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.474739] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-58c14eeb-fc62-43f9-85d8-3e091c80fc50 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277598', 'volume_id': 'cd524289-5d08-4b7f-a083-259e501f1fc1', 'name': 'volume-cd524289-5d08-4b7f-a083-259e501f1fc1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd03c47da-7e73-4474-ad9b-07d27629dcae', 'attached_at': '', 'detached_at': '', 'volume_id': 'cd524289-5d08-4b7f-a083-259e501f1fc1', 'serial': 'cd524289-5d08-4b7f-a083-259e501f1fc1'} {{(pid=63021) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1142.815937] env[63021]: DEBUG nova.scheduler.client.report [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1143.019185] env[63021]: DEBUG nova.objects.instance [None req-58c14eeb-fc62-43f9-85d8-3e091c80fc50 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lazy-loading 'flavor' on Instance uuid d03c47da-7e73-4474-ad9b-07d27629dcae {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1143.826067] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.632s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.026848] env[63021]: DEBUG oslo_concurrency.lockutils [None req-58c14eeb-fc62-43f9-85d8-3e091c80fc50 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "d03c47da-7e73-4474-ad9b-07d27629dcae" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.279s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1145.014765] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "d03c47da-7e73-4474-ad9b-07d27629dcae" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1145.015190] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "d03c47da-7e73-4474-ad9b-07d27629dcae" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1145.015275] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "d03c47da-7e73-4474-ad9b-07d27629dcae-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1145.015478] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "d03c47da-7e73-4474-ad9b-07d27629dcae-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1145.015645] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "d03c47da-7e73-4474-ad9b-07d27629dcae-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1145.018458] env[63021]: INFO nova.compute.manager [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Terminating instance [ 1145.020007] env[63021]: DEBUG nova.compute.manager [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1145.020217] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1145.021058] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efa940e8-3fd7-480e-885d-119fc989baa1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.028605] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1145.028845] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5545f590-5ba4-4b16-b243-e9e22a1d7e01 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.034603] env[63021]: DEBUG oslo_vmware.api [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1145.034603] env[63021]: value = "task-1294374" [ 1145.034603] env[63021]: _type = "Task" [ 1145.034603] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.042639] env[63021]: DEBUG oslo_vmware.api [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294374, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.365912] env[63021]: INFO nova.compute.manager [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Swapping old allocation on dict_keys(['6ef6c041-2b0d-4bc3-8662-5d679a77a86d']) held by migration ff919176-bc4f-4549-8596-4a1f9383b098 for instance [ 1145.386265] env[63021]: DEBUG nova.scheduler.client.report [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Overwriting current allocation {'allocations': {'6ef6c041-2b0d-4bc3-8662-5d679a77a86d': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 139}}, 'project_id': '5af87713ae6b477ca8a584aed287889d', 'user_id': '31c5d292015043aa8bfc6647e7405ea3', 'consumer_generation': 1} on consumer 4041dfbf-fa94-4ff8-9176-e7088cdf241f {{(pid=63021) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1145.545550] env[63021]: DEBUG oslo_vmware.api [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294374, 'name': PowerOffVM_Task, 'duration_secs': 0.166524} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.545550] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1145.545550] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1145.545818] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2f914942-8270-4f22-8dbf-9959ea7b1c46 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.565099] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "refresh_cache-4041dfbf-fa94-4ff8-9176-e7088cdf241f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1145.565280] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquired lock "refresh_cache-4041dfbf-fa94-4ff8-9176-e7088cdf241f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1145.565454] env[63021]: DEBUG nova.network.neutron [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1145.606272] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1145.606494] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Deleting contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1145.606672] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Deleting the datastore file [datastore2] d03c47da-7e73-4474-ad9b-07d27629dcae {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1145.606936] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a798b616-683c-4556-b2bf-3e4b88844e6f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.614246] env[63021]: DEBUG oslo_vmware.api [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1145.614246] env[63021]: value = "task-1294376" [ 1145.614246] env[63021]: _type = "Task" [ 1145.614246] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.621355] env[63021]: DEBUG oslo_vmware.api [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294376, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.124467] env[63021]: DEBUG oslo_vmware.api [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294376, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.127057} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.124852] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1146.124898] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Deleted contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1146.125089] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1146.125273] env[63021]: INFO nova.compute.manager [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1146.125514] env[63021]: DEBUG oslo.service.loopingcall [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1146.125708] env[63021]: DEBUG nova.compute.manager [-] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1146.125801] env[63021]: DEBUG nova.network.neutron [-] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1146.390158] env[63021]: DEBUG nova.network.neutron [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Updating instance_info_cache with network_info: [{"id": "b9d1b591-e9bc-4e70-9c76-1a0673bca1d5", "address": "fa:16:3e:1b:b0:e6", "network": {"id": "9502e364-60e1-47ef-a1f7-f3e3ff45cf69", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-164717784-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5af87713ae6b477ca8a584aed287889d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9d1b591-e9", "ovs_interfaceid": "b9d1b591-e9bc-4e70-9c76-1a0673bca1d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1146.875265] env[63021]: DEBUG nova.compute.manager [req-10802261-ca82-43b1-90f9-0ec2309e5a64 req-ad986d54-d4f3-47f8-bb17-fcb07ff32e90 service nova] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Received event network-vif-deleted-43370cd4-afaf-437d-8079-cf9ad90682aa {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1146.875478] env[63021]: INFO nova.compute.manager [req-10802261-ca82-43b1-90f9-0ec2309e5a64 req-ad986d54-d4f3-47f8-bb17-fcb07ff32e90 service nova] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Neutron deleted interface 43370cd4-afaf-437d-8079-cf9ad90682aa; detaching it from the instance and deleting it from the info cache [ 1146.875683] env[63021]: DEBUG nova.network.neutron [req-10802261-ca82-43b1-90f9-0ec2309e5a64 req-ad986d54-d4f3-47f8-bb17-fcb07ff32e90 service nova] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1146.893526] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Releasing lock "refresh_cache-4041dfbf-fa94-4ff8-9176-e7088cdf241f" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1146.893637] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1146.894590] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e3b1675f-68b0-458d-93aa-ae7fee38ba4a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.901351] env[63021]: DEBUG oslo_vmware.api [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1146.901351] env[63021]: value = "task-1294377" [ 1146.901351] env[63021]: _type = "Task" [ 1146.901351] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.909677] env[63021]: DEBUG oslo_vmware.api [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294377, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.316856] env[63021]: DEBUG nova.network.neutron [-] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1147.378174] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bb697077-f405-4d95-8c0b-c0a7fd8e96b5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.388247] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01d49ad8-f9a9-45d4-827f-f9aeabc0f854 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.409949] env[63021]: DEBUG oslo_vmware.api [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294377, 'name': PowerOffVM_Task, 'duration_secs': 0.158602} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.415495] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1147.416217] env[63021]: DEBUG nova.virt.hardware [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1147.416468] env[63021]: DEBUG nova.virt.hardware [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1147.416628] env[63021]: DEBUG nova.virt.hardware [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1147.416821] env[63021]: DEBUG nova.virt.hardware [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1147.416957] env[63021]: DEBUG nova.virt.hardware [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1147.417124] env[63021]: DEBUG nova.virt.hardware [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1147.417328] env[63021]: DEBUG nova.virt.hardware [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1147.417484] env[63021]: DEBUG nova.virt.hardware [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1147.417676] env[63021]: DEBUG nova.virt.hardware [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1147.418407] env[63021]: DEBUG nova.virt.hardware [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1147.418590] env[63021]: DEBUG nova.virt.hardware [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1147.423851] env[63021]: DEBUG nova.compute.manager [req-10802261-ca82-43b1-90f9-0ec2309e5a64 req-ad986d54-d4f3-47f8-bb17-fcb07ff32e90 service nova] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Detach interface failed, port_id=43370cd4-afaf-437d-8079-cf9ad90682aa, reason: Instance d03c47da-7e73-4474-ad9b-07d27629dcae could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1147.424048] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c7be4b5b-1699-4e46-a01a-9f4df7f51acc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.438668] env[63021]: DEBUG oslo_vmware.api [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1147.438668] env[63021]: value = "task-1294378" [ 1147.438668] env[63021]: _type = "Task" [ 1147.438668] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.446359] env[63021]: DEBUG oslo_vmware.api [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294378, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.819781] env[63021]: INFO nova.compute.manager [-] [instance: d03c47da-7e73-4474-ad9b-07d27629dcae] Took 1.69 seconds to deallocate network for instance. [ 1147.948715] env[63021]: DEBUG oslo_vmware.api [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294378, 'name': ReconfigVM_Task, 'duration_secs': 0.133146} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.949529] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a05945f7-b45d-4759-beee-a7dcf6203dd9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.967558] env[63021]: DEBUG nova.virt.hardware [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1147.967853] env[63021]: DEBUG nova.virt.hardware [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1147.968039] env[63021]: DEBUG nova.virt.hardware [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1147.968244] env[63021]: DEBUG nova.virt.hardware [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1147.968396] env[63021]: DEBUG nova.virt.hardware [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1147.968549] env[63021]: DEBUG nova.virt.hardware [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1147.968782] env[63021]: DEBUG nova.virt.hardware [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1147.968963] env[63021]: DEBUG nova.virt.hardware [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1147.969134] env[63021]: DEBUG nova.virt.hardware [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1147.969299] env[63021]: DEBUG nova.virt.hardware [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1147.969474] env[63021]: DEBUG nova.virt.hardware [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1147.970291] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc3db0a4-572d-4c23-aaa2-60801131d3dc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.975831] env[63021]: DEBUG oslo_vmware.api [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1147.975831] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52ca6295-605b-c718-e07d-eec395e08ce4" [ 1147.975831] env[63021]: _type = "Task" [ 1147.975831] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.984206] env[63021]: DEBUG oslo_vmware.api [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52ca6295-605b-c718-e07d-eec395e08ce4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.326211] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1148.326610] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1148.326749] env[63021]: DEBUG nova.objects.instance [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lazy-loading 'resources' on Instance uuid d03c47da-7e73-4474-ad9b-07d27629dcae {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1148.486438] env[63021]: DEBUG oslo_vmware.api [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52ca6295-605b-c718-e07d-eec395e08ce4, 'name': SearchDatastore_Task, 'duration_secs': 0.008444} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.491786] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Reconfiguring VM instance instance-00000069 to detach disk 2000 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1148.492072] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2b1e6c99-e187-441c-b331-569e2005f8fa {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.509759] env[63021]: DEBUG oslo_vmware.api [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1148.509759] env[63021]: value = "task-1294379" [ 1148.509759] env[63021]: _type = "Task" [ 1148.509759] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.517188] env[63021]: DEBUG oslo_vmware.api [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294379, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.872157] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45725a44-aa8e-41c3-aa5a-24e232000171 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.879678] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16ffa230-676c-46c5-8a5d-8a31ea3af1a2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.909714] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b89671a-2882-4f67-b6d2-8e456b95e0b1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.917658] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b15798f7-d913-4ab4-a0ac-78364e8b8add {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.930263] env[63021]: DEBUG nova.compute.provider_tree [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1149.019260] env[63021]: DEBUG oslo_vmware.api [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294379, 'name': ReconfigVM_Task, 'duration_secs': 0.174885} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.019535] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Reconfigured VM instance instance-00000069 to detach disk 2000 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1149.020313] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0280024f-ccd2-4ad7-b194-4559b5e37dc7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.041520] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] 4041dfbf-fa94-4ff8-9176-e7088cdf241f/4041dfbf-fa94-4ff8-9176-e7088cdf241f.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1149.041786] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d666908f-2b45-44b6-bea3-27c0dbaa99ff {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.059047] env[63021]: DEBUG oslo_vmware.api [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1149.059047] env[63021]: value = "task-1294380" [ 1149.059047] env[63021]: _type = "Task" [ 1149.059047] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.066259] env[63021]: DEBUG oslo_vmware.api [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294380, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.432965] env[63021]: DEBUG nova.scheduler.client.report [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1149.569601] env[63021]: DEBUG oslo_vmware.api [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294380, 'name': ReconfigVM_Task, 'duration_secs': 0.255748} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.569885] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Reconfigured VM instance instance-00000069 to attach disk [datastore2] 4041dfbf-fa94-4ff8-9176-e7088cdf241f/4041dfbf-fa94-4ff8-9176-e7088cdf241f.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1149.570683] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5808e62a-73f5-4a99-a3f3-7a0880c0bc3f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.588649] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53abbf8c-d5e2-43b2-9530-6f068bdff2ba {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.605422] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d33d048-e811-4aba-8f36-1fb750714d98 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.623375] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-209e7ae8-64ad-4230-9021-774ce3ebcd90 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.629463] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1149.629673] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0d8ee0a5-3492-41ee-b60c-8f154f9d3904 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.635315] env[63021]: DEBUG oslo_vmware.api [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1149.635315] env[63021]: value = "task-1294381" [ 1149.635315] env[63021]: _type = "Task" [ 1149.635315] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.642011] env[63021]: DEBUG oslo_vmware.api [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294381, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.938027] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.611s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1149.962853] env[63021]: INFO nova.scheduler.client.report [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Deleted allocations for instance d03c47da-7e73-4474-ad9b-07d27629dcae [ 1150.144897] env[63021]: DEBUG oslo_vmware.api [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294381, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.470743] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b4a6cd20-b806-4aa6-b7c6-68b263c62477 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "d03c47da-7e73-4474-ad9b-07d27629dcae" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.456s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1150.599389] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1150.599572] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63021) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1150.645109] env[63021]: DEBUG oslo_vmware.api [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294381, 'name': PowerOnVM_Task, 'duration_secs': 0.561822} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.645383] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1151.655378] env[63021]: INFO nova.compute.manager [None req-c4711be8-7037-4633-b04c-f1027bf40565 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Updating instance to original state: 'active' [ 1152.600597] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1153.007530] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "2e71d990-8bae-4ae8-b7c7-dd26984a08f5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1153.007841] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "2e71d990-8bae-4ae8-b7c7-dd26984a08f5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.257026] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1c9333b7-4933-43c7-b302-a6ae296c8826 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "4041dfbf-fa94-4ff8-9176-e7088cdf241f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1153.257334] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1c9333b7-4933-43c7-b302-a6ae296c8826 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "4041dfbf-fa94-4ff8-9176-e7088cdf241f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.257568] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1c9333b7-4933-43c7-b302-a6ae296c8826 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "4041dfbf-fa94-4ff8-9176-e7088cdf241f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1153.257855] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1c9333b7-4933-43c7-b302-a6ae296c8826 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "4041dfbf-fa94-4ff8-9176-e7088cdf241f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.258048] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1c9333b7-4933-43c7-b302-a6ae296c8826 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "4041dfbf-fa94-4ff8-9176-e7088cdf241f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.260252] env[63021]: INFO nova.compute.manager [None req-1c9333b7-4933-43c7-b302-a6ae296c8826 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Terminating instance [ 1153.262127] env[63021]: DEBUG nova.compute.manager [None req-1c9333b7-4933-43c7-b302-a6ae296c8826 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1153.262327] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-1c9333b7-4933-43c7-b302-a6ae296c8826 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1153.263173] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8b5769d-746f-42ba-acd2-8cdae5b492dc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.270578] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c9333b7-4933-43c7-b302-a6ae296c8826 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1153.270803] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-613dc955-84dd-4066-8920-af5bc550baca {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.276381] env[63021]: DEBUG oslo_vmware.api [None req-1c9333b7-4933-43c7-b302-a6ae296c8826 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1153.276381] env[63021]: value = "task-1294382" [ 1153.276381] env[63021]: _type = "Task" [ 1153.276381] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.284088] env[63021]: DEBUG oslo_vmware.api [None req-1c9333b7-4933-43c7-b302-a6ae296c8826 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294382, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.511123] env[63021]: DEBUG nova.compute.manager [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1153.595369] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1153.786038] env[63021]: DEBUG oslo_vmware.api [None req-1c9333b7-4933-43c7-b302-a6ae296c8826 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294382, 'name': PowerOffVM_Task, 'duration_secs': 0.199712} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.786342] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c9333b7-4933-43c7-b302-a6ae296c8826 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1153.786520] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-1c9333b7-4933-43c7-b302-a6ae296c8826 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1153.786771] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-417c703c-bf02-4256-a76c-d152d8beb180 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.895010] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-1c9333b7-4933-43c7-b302-a6ae296c8826 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1153.895262] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-1c9333b7-4933-43c7-b302-a6ae296c8826 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Deleting contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1153.895503] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c9333b7-4933-43c7-b302-a6ae296c8826 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Deleting the datastore file [datastore2] 4041dfbf-fa94-4ff8-9176-e7088cdf241f {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1153.898948] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8514b7a0-603e-407e-8943-396ae5f35519 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.902367] env[63021]: DEBUG oslo_vmware.api [None req-1c9333b7-4933-43c7-b302-a6ae296c8826 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1153.902367] env[63021]: value = "task-1294384" [ 1153.902367] env[63021]: _type = "Task" [ 1153.902367] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.910390] env[63021]: DEBUG oslo_vmware.api [None req-1c9333b7-4933-43c7-b302-a6ae296c8826 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294384, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.034131] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1154.034442] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1154.036125] env[63021]: INFO nova.compute.claims [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1154.412495] env[63021]: DEBUG oslo_vmware.api [None req-1c9333b7-4933-43c7-b302-a6ae296c8826 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294384, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.166841} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.412707] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c9333b7-4933-43c7-b302-a6ae296c8826 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1154.412896] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-1c9333b7-4933-43c7-b302-a6ae296c8826 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Deleted contents of the VM from datastore datastore2 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1154.413087] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-1c9333b7-4933-43c7-b302-a6ae296c8826 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1154.413265] env[63021]: INFO nova.compute.manager [None req-1c9333b7-4933-43c7-b302-a6ae296c8826 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1154.413523] env[63021]: DEBUG oslo.service.loopingcall [None req-1c9333b7-4933-43c7-b302-a6ae296c8826 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1154.413742] env[63021]: DEBUG nova.compute.manager [-] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1154.413843] env[63021]: DEBUG nova.network.neutron [-] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1154.599221] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1154.599421] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Starting heal instance info cache {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1154.599557] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Rebuilding the list of instances to heal {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1154.883688] env[63021]: DEBUG nova.compute.manager [req-9b4bbcce-bbb2-4ada-99b0-3f21740dbe21 req-40f66a6c-e4a1-49fa-9c21-2cb90717c267 service nova] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Received event network-vif-deleted-b9d1b591-e9bc-4e70-9c76-1a0673bca1d5 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1154.883870] env[63021]: INFO nova.compute.manager [req-9b4bbcce-bbb2-4ada-99b0-3f21740dbe21 req-40f66a6c-e4a1-49fa-9c21-2cb90717c267 service nova] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Neutron deleted interface b9d1b591-e9bc-4e70-9c76-1a0673bca1d5; detaching it from the instance and deleting it from the info cache [ 1154.884108] env[63021]: DEBUG nova.network.neutron [req-9b4bbcce-bbb2-4ada-99b0-3f21740dbe21 req-40f66a6c-e4a1-49fa-9c21-2cb90717c267 service nova] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1155.085123] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2cebb63-ad73-4486-8529-3cee5ee8e950 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.092714] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3cb3daf-5331-4f08-996a-27bfe367623b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.123765] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Skipping network cache update for instance because it is being deleted. {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 1155.123958] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Skipping network cache update for instance because it is Building. {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1155.124057] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Didn't find any instances for network info cache update. {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1155.124481] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1155.125227] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bd1d89b-dbc2-431b-b9d9-c0867d79ed0e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.127937] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1155.128130] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1155.133813] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a71b816-c13e-4100-9975-2363a619d32d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.147957] env[63021]: DEBUG nova.compute.provider_tree [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1155.332422] env[63021]: DEBUG nova.network.neutron [-] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1155.386782] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-369733e1-4370-4e78-a248-0f9133e216cb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.397115] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d700be5-ff4a-4f4c-bcd9-ae4eae683056 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.418309] env[63021]: DEBUG nova.compute.manager [req-9b4bbcce-bbb2-4ada-99b0-3f21740dbe21 req-40f66a6c-e4a1-49fa-9c21-2cb90717c267 service nova] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Detach interface failed, port_id=b9d1b591-e9bc-4e70-9c76-1a0673bca1d5, reason: Instance 4041dfbf-fa94-4ff8-9176-e7088cdf241f could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1155.631188] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.650441] env[63021]: DEBUG nova.scheduler.client.report [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1155.835529] env[63021]: INFO nova.compute.manager [-] [instance: 4041dfbf-fa94-4ff8-9176-e7088cdf241f] Took 1.42 seconds to deallocate network for instance. [ 1156.155222] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.121s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1156.155759] env[63021]: DEBUG nova.compute.manager [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1156.159865] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.529s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1156.160066] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1156.160236] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63021) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1156.161274] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e229b172-8eb0-4aea-9ffb-2575c054681b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.169297] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f09ead1e-9d9e-4fad-8769-bc987b8ab83f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.182737] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ef66a7d-7778-4327-89db-f91a88026f3c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.189051] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f726e88f-8e4d-4746-8626-c16fc468b33e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.217796] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180041MB free_disk=139GB free_vcpus=48 pci_devices=None {{(pid=63021) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1156.217954] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1156.218173] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1156.342270] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1c9333b7-4933-43c7-b302-a6ae296c8826 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1156.662780] env[63021]: DEBUG nova.compute.utils [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1156.664178] env[63021]: DEBUG nova.compute.manager [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1156.664347] env[63021]: DEBUG nova.network.neutron [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1156.710682] env[63021]: DEBUG nova.policy [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5431727705ff4c5483cd9a5b9cd104f6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '547d9a3099b94ff488ef6929ae591bac', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 1156.970995] env[63021]: DEBUG nova.network.neutron [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Successfully created port: e9488dcd-9157-488c-8f45-8c497d12ea22 {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1157.166980] env[63021]: DEBUG nova.compute.manager [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1157.240822] env[63021]: WARNING nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 4041dfbf-fa94-4ff8-9176-e7088cdf241f is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1157.240992] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 2e71d990-8bae-4ae8-b7c7-dd26984a08f5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1157.241182] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=63021) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1157.241324] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=63021) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1157.277470] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa712c89-02e6-42e1-93de-b995ddbaf638 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.285168] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d801ffe-b9d5-4bc9-8855-8ba670ccaf96 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.315180] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd88c819-3ccb-4b91-aef1-7376705fb062 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.321654] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5f93b98-3ce3-481d-a361-14f39619acc8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.334612] env[63021]: DEBUG nova.compute.provider_tree [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1157.837542] env[63021]: DEBUG nova.scheduler.client.report [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1158.176557] env[63021]: DEBUG nova.compute.manager [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1158.200882] env[63021]: DEBUG nova.virt.hardware [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1158.201158] env[63021]: DEBUG nova.virt.hardware [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1158.201330] env[63021]: DEBUG nova.virt.hardware [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1158.201517] env[63021]: DEBUG nova.virt.hardware [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1158.203441] env[63021]: DEBUG nova.virt.hardware [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1158.203441] env[63021]: DEBUG nova.virt.hardware [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1158.203441] env[63021]: DEBUG nova.virt.hardware [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1158.203441] env[63021]: DEBUG nova.virt.hardware [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1158.203441] env[63021]: DEBUG nova.virt.hardware [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1158.203441] env[63021]: DEBUG nova.virt.hardware [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1158.203441] env[63021]: DEBUG nova.virt.hardware [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1158.204099] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cc3a2c7-e2ee-4ddf-b079-3d34d6ca8a8d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.211539] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-592f2a53-6b7d-4b14-898b-74f6f5843340 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.342151] env[63021]: DEBUG nova.compute.manager [req-712fedaa-5aa9-4a63-ad0e-4eb10d180726 req-e01b8c9c-0e8c-40fa-8077-bc80a834de2f service nova] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Received event network-vif-plugged-e9488dcd-9157-488c-8f45-8c497d12ea22 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1158.342384] env[63021]: DEBUG oslo_concurrency.lockutils [req-712fedaa-5aa9-4a63-ad0e-4eb10d180726 req-e01b8c9c-0e8c-40fa-8077-bc80a834de2f service nova] Acquiring lock "2e71d990-8bae-4ae8-b7c7-dd26984a08f5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1158.342596] env[63021]: DEBUG oslo_concurrency.lockutils [req-712fedaa-5aa9-4a63-ad0e-4eb10d180726 req-e01b8c9c-0e8c-40fa-8077-bc80a834de2f service nova] Lock "2e71d990-8bae-4ae8-b7c7-dd26984a08f5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1158.342779] env[63021]: DEBUG oslo_concurrency.lockutils [req-712fedaa-5aa9-4a63-ad0e-4eb10d180726 req-e01b8c9c-0e8c-40fa-8077-bc80a834de2f service nova] Lock "2e71d990-8bae-4ae8-b7c7-dd26984a08f5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1158.343022] env[63021]: DEBUG nova.compute.manager [req-712fedaa-5aa9-4a63-ad0e-4eb10d180726 req-e01b8c9c-0e8c-40fa-8077-bc80a834de2f service nova] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] No waiting events found dispatching network-vif-plugged-e9488dcd-9157-488c-8f45-8c497d12ea22 {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1158.343418] env[63021]: WARNING nova.compute.manager [req-712fedaa-5aa9-4a63-ad0e-4eb10d180726 req-e01b8c9c-0e8c-40fa-8077-bc80a834de2f service nova] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Received unexpected event network-vif-plugged-e9488dcd-9157-488c-8f45-8c497d12ea22 for instance with vm_state building and task_state spawning. [ 1158.344057] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63021) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1158.344232] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.126s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1158.344736] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1c9333b7-4933-43c7-b302-a6ae296c8826 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.003s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1158.344958] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1c9333b7-4933-43c7-b302-a6ae296c8826 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1158.361664] env[63021]: INFO nova.scheduler.client.report [None req-1c9333b7-4933-43c7-b302-a6ae296c8826 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Deleted allocations for instance 4041dfbf-fa94-4ff8-9176-e7088cdf241f [ 1158.426057] env[63021]: DEBUG nova.network.neutron [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Successfully updated port: e9488dcd-9157-488c-8f45-8c497d12ea22 {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1158.869364] env[63021]: DEBUG oslo_concurrency.lockutils [None req-1c9333b7-4933-43c7-b302-a6ae296c8826 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "4041dfbf-fa94-4ff8-9176-e7088cdf241f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.612s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1158.929304] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "refresh_cache-2e71d990-8bae-4ae8-b7c7-dd26984a08f5" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1158.929304] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquired lock "refresh_cache-2e71d990-8bae-4ae8-b7c7-dd26984a08f5" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1158.929525] env[63021]: DEBUG nova.network.neutron [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1159.460275] env[63021]: DEBUG nova.network.neutron [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1159.587258] env[63021]: DEBUG nova.network.neutron [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Updating instance_info_cache with network_info: [{"id": "e9488dcd-9157-488c-8f45-8c497d12ea22", "address": "fa:16:3e:99:86:b2", "network": {"id": "d27bb3cc-fb80-4b03-9288-ed17a0d05140", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-122910456-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "547d9a3099b94ff488ef6929ae591bac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1c797172-a569-458e-aeb0-3f21e589a740", "external-id": "nsx-vlan-transportzone-957", "segmentation_id": 957, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9488dcd-91", "ovs_interfaceid": "e9488dcd-9157-488c-8f45-8c497d12ea22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1159.821887] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1159.822118] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1160.089779] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Releasing lock "refresh_cache-2e71d990-8bae-4ae8-b7c7-dd26984a08f5" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1160.090185] env[63021]: DEBUG nova.compute.manager [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Instance network_info: |[{"id": "e9488dcd-9157-488c-8f45-8c497d12ea22", "address": "fa:16:3e:99:86:b2", "network": {"id": "d27bb3cc-fb80-4b03-9288-ed17a0d05140", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-122910456-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "547d9a3099b94ff488ef6929ae591bac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1c797172-a569-458e-aeb0-3f21e589a740", "external-id": "nsx-vlan-transportzone-957", "segmentation_id": 957, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9488dcd-91", "ovs_interfaceid": "e9488dcd-9157-488c-8f45-8c497d12ea22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1160.090632] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:99:86:b2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1c797172-a569-458e-aeb0-3f21e589a740', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e9488dcd-9157-488c-8f45-8c497d12ea22', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1160.098652] env[63021]: DEBUG oslo.service.loopingcall [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1160.098976] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1160.099344] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-70d4c86a-c74f-49a6-9832-eda1d7349ba6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.120675] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1160.120675] env[63021]: value = "task-1294385" [ 1160.120675] env[63021]: _type = "Task" [ 1160.120675] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.127927] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294385, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.218559] env[63021]: DEBUG oslo_concurrency.lockutils [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "e87e4421-d0a9-4d5d-816e-abba7caa687e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1160.218796] env[63021]: DEBUG oslo_concurrency.lockutils [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "e87e4421-d0a9-4d5d-816e-abba7caa687e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1160.378343] env[63021]: DEBUG nova.compute.manager [req-1f1aea59-7032-432c-9cdc-ed6b91a19ba1 req-2f0b2392-8b60-4d30-b953-4af0d8daf047 service nova] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Received event network-changed-e9488dcd-9157-488c-8f45-8c497d12ea22 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1160.378549] env[63021]: DEBUG nova.compute.manager [req-1f1aea59-7032-432c-9cdc-ed6b91a19ba1 req-2f0b2392-8b60-4d30-b953-4af0d8daf047 service nova] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Refreshing instance network info cache due to event network-changed-e9488dcd-9157-488c-8f45-8c497d12ea22. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1160.378758] env[63021]: DEBUG oslo_concurrency.lockutils [req-1f1aea59-7032-432c-9cdc-ed6b91a19ba1 req-2f0b2392-8b60-4d30-b953-4af0d8daf047 service nova] Acquiring lock "refresh_cache-2e71d990-8bae-4ae8-b7c7-dd26984a08f5" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1160.378930] env[63021]: DEBUG oslo_concurrency.lockutils [req-1f1aea59-7032-432c-9cdc-ed6b91a19ba1 req-2f0b2392-8b60-4d30-b953-4af0d8daf047 service nova] Acquired lock "refresh_cache-2e71d990-8bae-4ae8-b7c7-dd26984a08f5" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1160.379129] env[63021]: DEBUG nova.network.neutron [req-1f1aea59-7032-432c-9cdc-ed6b91a19ba1 req-2f0b2392-8b60-4d30-b953-4af0d8daf047 service nova] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Refreshing network info cache for port e9488dcd-9157-488c-8f45-8c497d12ea22 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1160.629830] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294385, 'name': CreateVM_Task, 'duration_secs': 0.300192} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.630182] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1160.630648] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1160.630819] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1160.631167] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1160.631407] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-562487e6-c109-4cf6-9897-93c4d637ef5b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.635459] env[63021]: DEBUG oslo_vmware.api [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1160.635459] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52f3ebc4-59a0-c77f-dc32-b66a46804076" [ 1160.635459] env[63021]: _type = "Task" [ 1160.635459] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.642890] env[63021]: DEBUG oslo_vmware.api [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52f3ebc4-59a0-c77f-dc32-b66a46804076, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.720934] env[63021]: DEBUG nova.compute.manager [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1161.084227] env[63021]: DEBUG nova.network.neutron [req-1f1aea59-7032-432c-9cdc-ed6b91a19ba1 req-2f0b2392-8b60-4d30-b953-4af0d8daf047 service nova] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Updated VIF entry in instance network info cache for port e9488dcd-9157-488c-8f45-8c497d12ea22. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1161.084617] env[63021]: DEBUG nova.network.neutron [req-1f1aea59-7032-432c-9cdc-ed6b91a19ba1 req-2f0b2392-8b60-4d30-b953-4af0d8daf047 service nova] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Updating instance_info_cache with network_info: [{"id": "e9488dcd-9157-488c-8f45-8c497d12ea22", "address": "fa:16:3e:99:86:b2", "network": {"id": "d27bb3cc-fb80-4b03-9288-ed17a0d05140", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-122910456-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "547d9a3099b94ff488ef6929ae591bac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1c797172-a569-458e-aeb0-3f21e589a740", "external-id": "nsx-vlan-transportzone-957", "segmentation_id": 957, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9488dcd-91", "ovs_interfaceid": "e9488dcd-9157-488c-8f45-8c497d12ea22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1161.146762] env[63021]: DEBUG oslo_vmware.api [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52f3ebc4-59a0-c77f-dc32-b66a46804076, 'name': SearchDatastore_Task, 'duration_secs': 0.009583} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.147132] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1161.147407] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1161.147708] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1161.147862] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1161.148067] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1161.148336] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b1c83100-6f58-4270-9f69-e23fec57ebb4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.156832] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1161.157013] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1161.157720] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c6e20f7-c3c5-4771-bd92-4fb356a70a49 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.162748] env[63021]: DEBUG oslo_vmware.api [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1161.162748] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5211b203-7708-b69b-4476-912007b6a373" [ 1161.162748] env[63021]: _type = "Task" [ 1161.162748] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.169767] env[63021]: DEBUG oslo_vmware.api [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5211b203-7708-b69b-4476-912007b6a373, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.242278] env[63021]: DEBUG oslo_concurrency.lockutils [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1161.242527] env[63021]: DEBUG oslo_concurrency.lockutils [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1161.244088] env[63021]: INFO nova.compute.claims [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1161.587877] env[63021]: DEBUG oslo_concurrency.lockutils [req-1f1aea59-7032-432c-9cdc-ed6b91a19ba1 req-2f0b2392-8b60-4d30-b953-4af0d8daf047 service nova] Releasing lock "refresh_cache-2e71d990-8bae-4ae8-b7c7-dd26984a08f5" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1161.673070] env[63021]: DEBUG oslo_vmware.api [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5211b203-7708-b69b-4476-912007b6a373, 'name': SearchDatastore_Task, 'duration_secs': 0.008844} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.673816] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c446321c-3d19-4fb3-9a41-bc7a7116334e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.678455] env[63021]: DEBUG oslo_vmware.api [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1161.678455] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]526a54f0-00f0-6422-9285-6db71e49d8ab" [ 1161.678455] env[63021]: _type = "Task" [ 1161.678455] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.685549] env[63021]: DEBUG oslo_vmware.api [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]526a54f0-00f0-6422-9285-6db71e49d8ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.189904] env[63021]: DEBUG oslo_vmware.api [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]526a54f0-00f0-6422-9285-6db71e49d8ab, 'name': SearchDatastore_Task, 'duration_secs': 0.030432} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.189904] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1162.189904] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 2e71d990-8bae-4ae8-b7c7-dd26984a08f5/2e71d990-8bae-4ae8-b7c7-dd26984a08f5.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1162.190163] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f8631de1-0ce8-4186-8131-2679f15d924e {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.196116] env[63021]: DEBUG oslo_vmware.api [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1162.196116] env[63021]: value = "task-1294386" [ 1162.196116] env[63021]: _type = "Task" [ 1162.196116] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.203327] env[63021]: DEBUG oslo_vmware.api [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294386, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.291155] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b6e4bff-c1b4-4c8e-9042-2e83040789d6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.298675] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c9827ad-e555-47f5-bc54-9e3ddc4114d7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.329366] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce1021ee-d30f-4692-8258-05617e519f6d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.336139] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9977bfa7-cba4-47b5-b721-700af97894f2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.349173] env[63021]: DEBUG nova.compute.provider_tree [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1162.705594] env[63021]: DEBUG oslo_vmware.api [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294386, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.851940] env[63021]: DEBUG nova.scheduler.client.report [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1163.206310] env[63021]: DEBUG oslo_vmware.api [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294386, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.512195} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.206589] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 2e71d990-8bae-4ae8-b7c7-dd26984a08f5/2e71d990-8bae-4ae8-b7c7-dd26984a08f5.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1163.206806] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1163.207078] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fb497b6f-4789-4d14-956b-df16dc3a0c51 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.214179] env[63021]: DEBUG oslo_vmware.api [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1163.214179] env[63021]: value = "task-1294387" [ 1163.214179] env[63021]: _type = "Task" [ 1163.214179] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.221486] env[63021]: DEBUG oslo_vmware.api [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294387, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.357091] env[63021]: DEBUG oslo_concurrency.lockutils [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.114s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1163.357224] env[63021]: DEBUG nova.compute.manager [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1163.723243] env[63021]: DEBUG oslo_vmware.api [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294387, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076323} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.723605] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1163.724250] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97e20785-bb2b-4029-8ead-98ecf5d77994 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.745276] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] 2e71d990-8bae-4ae8-b7c7-dd26984a08f5/2e71d990-8bae-4ae8-b7c7-dd26984a08f5.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1163.745516] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-92d4c2fd-3a0f-4a08-811a-b76a5487cc67 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.763998] env[63021]: DEBUG oslo_vmware.api [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1163.763998] env[63021]: value = "task-1294388" [ 1163.763998] env[63021]: _type = "Task" [ 1163.763998] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.771151] env[63021]: DEBUG oslo_vmware.api [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294388, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.861856] env[63021]: DEBUG nova.compute.utils [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1163.863278] env[63021]: DEBUG nova.compute.manager [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1163.863452] env[63021]: DEBUG nova.network.neutron [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1163.901706] env[63021]: DEBUG nova.policy [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '31c5d292015043aa8bfc6647e7405ea3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5af87713ae6b477ca8a584aed287889d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 1164.157097] env[63021]: DEBUG nova.network.neutron [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Successfully created port: 5298015f-1096-4db7-8e70-1c91859e685f {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1164.274591] env[63021]: DEBUG oslo_vmware.api [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294388, 'name': ReconfigVM_Task, 'duration_secs': 0.253889} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.274869] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Reconfigured VM instance instance-0000006d to attach disk [datastore1] 2e71d990-8bae-4ae8-b7c7-dd26984a08f5/2e71d990-8bae-4ae8-b7c7-dd26984a08f5.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1164.275452] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5a9b4b48-5cc1-4ded-a128-95cc868fbcb2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.281452] env[63021]: DEBUG oslo_vmware.api [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1164.281452] env[63021]: value = "task-1294389" [ 1164.281452] env[63021]: _type = "Task" [ 1164.281452] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.289046] env[63021]: DEBUG oslo_vmware.api [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294389, 'name': Rename_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.367057] env[63021]: DEBUG nova.compute.manager [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1164.791247] env[63021]: DEBUG oslo_vmware.api [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294389, 'name': Rename_Task, 'duration_secs': 0.144329} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.791530] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1164.791761] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d54d88e4-9f5c-4a90-a0ad-c226d3431f72 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.797513] env[63021]: DEBUG oslo_vmware.api [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1164.797513] env[63021]: value = "task-1294390" [ 1164.797513] env[63021]: _type = "Task" [ 1164.797513] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.804596] env[63021]: DEBUG oslo_vmware.api [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294390, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.307268] env[63021]: DEBUG oslo_vmware.api [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294390, 'name': PowerOnVM_Task, 'duration_secs': 0.474776} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.307540] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1165.307773] env[63021]: INFO nova.compute.manager [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Took 7.13 seconds to spawn the instance on the hypervisor. [ 1165.307974] env[63021]: DEBUG nova.compute.manager [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1165.308784] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b27233a6-c76b-4a94-9e42-1ed26c802662 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.376906] env[63021]: DEBUG nova.compute.manager [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1165.402476] env[63021]: DEBUG nova.virt.hardware [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1165.402680] env[63021]: DEBUG nova.virt.hardware [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1165.402839] env[63021]: DEBUG nova.virt.hardware [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1165.403056] env[63021]: DEBUG nova.virt.hardware [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1165.403238] env[63021]: DEBUG nova.virt.hardware [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1165.403388] env[63021]: DEBUG nova.virt.hardware [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1165.403601] env[63021]: DEBUG nova.virt.hardware [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1165.403762] env[63021]: DEBUG nova.virt.hardware [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1165.403934] env[63021]: DEBUG nova.virt.hardware [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1165.404117] env[63021]: DEBUG nova.virt.hardware [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1165.404299] env[63021]: DEBUG nova.virt.hardware [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1165.405211] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e0653a8-e620-4064-a668-c1939e6f743f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.413104] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef4eb9e6-e5a4-4b46-ac80-e18648d8f932 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.542036] env[63021]: DEBUG nova.compute.manager [req-2ff48091-db72-4dfb-b59d-3925a05229c4 req-4ff479e7-0cfa-41b7-9f16-276f1b97309f service nova] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Received event network-vif-plugged-5298015f-1096-4db7-8e70-1c91859e685f {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1165.542271] env[63021]: DEBUG oslo_concurrency.lockutils [req-2ff48091-db72-4dfb-b59d-3925a05229c4 req-4ff479e7-0cfa-41b7-9f16-276f1b97309f service nova] Acquiring lock "e87e4421-d0a9-4d5d-816e-abba7caa687e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.542479] env[63021]: DEBUG oslo_concurrency.lockutils [req-2ff48091-db72-4dfb-b59d-3925a05229c4 req-4ff479e7-0cfa-41b7-9f16-276f1b97309f service nova] Lock "e87e4421-d0a9-4d5d-816e-abba7caa687e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.542646] env[63021]: DEBUG oslo_concurrency.lockutils [req-2ff48091-db72-4dfb-b59d-3925a05229c4 req-4ff479e7-0cfa-41b7-9f16-276f1b97309f service nova] Lock "e87e4421-d0a9-4d5d-816e-abba7caa687e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.542815] env[63021]: DEBUG nova.compute.manager [req-2ff48091-db72-4dfb-b59d-3925a05229c4 req-4ff479e7-0cfa-41b7-9f16-276f1b97309f service nova] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] No waiting events found dispatching network-vif-plugged-5298015f-1096-4db7-8e70-1c91859e685f {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1165.542985] env[63021]: WARNING nova.compute.manager [req-2ff48091-db72-4dfb-b59d-3925a05229c4 req-4ff479e7-0cfa-41b7-9f16-276f1b97309f service nova] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Received unexpected event network-vif-plugged-5298015f-1096-4db7-8e70-1c91859e685f for instance with vm_state building and task_state spawning. [ 1165.629630] env[63021]: DEBUG nova.network.neutron [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Successfully updated port: 5298015f-1096-4db7-8e70-1c91859e685f {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1165.829961] env[63021]: INFO nova.compute.manager [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Took 11.81 seconds to build instance. [ 1166.043932] env[63021]: DEBUG nova.compute.manager [req-7397d874-bd88-4a4f-a818-16907fd6ddd5 req-18b2a2df-b7e9-4875-b88b-ca5e3ceec9de service nova] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Received event network-changed-e9488dcd-9157-488c-8f45-8c497d12ea22 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1166.044137] env[63021]: DEBUG nova.compute.manager [req-7397d874-bd88-4a4f-a818-16907fd6ddd5 req-18b2a2df-b7e9-4875-b88b-ca5e3ceec9de service nova] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Refreshing instance network info cache due to event network-changed-e9488dcd-9157-488c-8f45-8c497d12ea22. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1166.044375] env[63021]: DEBUG oslo_concurrency.lockutils [req-7397d874-bd88-4a4f-a818-16907fd6ddd5 req-18b2a2df-b7e9-4875-b88b-ca5e3ceec9de service nova] Acquiring lock "refresh_cache-2e71d990-8bae-4ae8-b7c7-dd26984a08f5" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1166.044619] env[63021]: DEBUG oslo_concurrency.lockutils [req-7397d874-bd88-4a4f-a818-16907fd6ddd5 req-18b2a2df-b7e9-4875-b88b-ca5e3ceec9de service nova] Acquired lock "refresh_cache-2e71d990-8bae-4ae8-b7c7-dd26984a08f5" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1166.044678] env[63021]: DEBUG nova.network.neutron [req-7397d874-bd88-4a4f-a818-16907fd6ddd5 req-18b2a2df-b7e9-4875-b88b-ca5e3ceec9de service nova] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Refreshing network info cache for port e9488dcd-9157-488c-8f45-8c497d12ea22 {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1166.134174] env[63021]: DEBUG oslo_concurrency.lockutils [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "refresh_cache-e87e4421-d0a9-4d5d-816e-abba7caa687e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1166.134174] env[63021]: DEBUG oslo_concurrency.lockutils [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquired lock "refresh_cache-e87e4421-d0a9-4d5d-816e-abba7caa687e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1166.134174] env[63021]: DEBUG nova.network.neutron [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1166.330944] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ce90d31f-1629-4838-b662-9e59622f16de tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "2e71d990-8bae-4ae8-b7c7-dd26984a08f5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.323s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1166.674999] env[63021]: DEBUG nova.network.neutron [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1166.845291] env[63021]: DEBUG nova.network.neutron [req-7397d874-bd88-4a4f-a818-16907fd6ddd5 req-18b2a2df-b7e9-4875-b88b-ca5e3ceec9de service nova] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Updated VIF entry in instance network info cache for port e9488dcd-9157-488c-8f45-8c497d12ea22. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1166.845969] env[63021]: DEBUG nova.network.neutron [req-7397d874-bd88-4a4f-a818-16907fd6ddd5 req-18b2a2df-b7e9-4875-b88b-ca5e3ceec9de service nova] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Updating instance_info_cache with network_info: [{"id": "e9488dcd-9157-488c-8f45-8c497d12ea22", "address": "fa:16:3e:99:86:b2", "network": {"id": "d27bb3cc-fb80-4b03-9288-ed17a0d05140", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-122910456-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "547d9a3099b94ff488ef6929ae591bac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1c797172-a569-458e-aeb0-3f21e589a740", "external-id": "nsx-vlan-transportzone-957", "segmentation_id": 957, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9488dcd-91", "ovs_interfaceid": "e9488dcd-9157-488c-8f45-8c497d12ea22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1166.869443] env[63021]: DEBUG nova.network.neutron [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Updating instance_info_cache with network_info: [{"id": "5298015f-1096-4db7-8e70-1c91859e685f", "address": "fa:16:3e:c8:03:c8", "network": {"id": "9502e364-60e1-47ef-a1f7-f3e3ff45cf69", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-164717784-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5af87713ae6b477ca8a584aed287889d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5298015f-10", "ovs_interfaceid": "5298015f-1096-4db7-8e70-1c91859e685f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1167.348599] env[63021]: DEBUG oslo_concurrency.lockutils [req-7397d874-bd88-4a4f-a818-16907fd6ddd5 req-18b2a2df-b7e9-4875-b88b-ca5e3ceec9de service nova] Releasing lock "refresh_cache-2e71d990-8bae-4ae8-b7c7-dd26984a08f5" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1167.372410] env[63021]: DEBUG oslo_concurrency.lockutils [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Releasing lock "refresh_cache-e87e4421-d0a9-4d5d-816e-abba7caa687e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1167.372764] env[63021]: DEBUG nova.compute.manager [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Instance network_info: |[{"id": "5298015f-1096-4db7-8e70-1c91859e685f", "address": "fa:16:3e:c8:03:c8", "network": {"id": "9502e364-60e1-47ef-a1f7-f3e3ff45cf69", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-164717784-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5af87713ae6b477ca8a584aed287889d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5298015f-10", "ovs_interfaceid": "5298015f-1096-4db7-8e70-1c91859e685f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1167.373202] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c8:03:c8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5298015f-1096-4db7-8e70-1c91859e685f', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1167.380952] env[63021]: DEBUG oslo.service.loopingcall [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1167.381201] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1167.381460] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-db2ecd83-640d-420c-907f-faaf272a9fbc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.402112] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1167.402112] env[63021]: value = "task-1294391" [ 1167.402112] env[63021]: _type = "Task" [ 1167.402112] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.409433] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294391, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.570058] env[63021]: DEBUG nova.compute.manager [req-894d1406-47a5-4752-a6b1-8692fdb2e800 req-16fc8cbf-d525-4a91-afba-bb8a11fc841e service nova] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Received event network-changed-5298015f-1096-4db7-8e70-1c91859e685f {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1167.570289] env[63021]: DEBUG nova.compute.manager [req-894d1406-47a5-4752-a6b1-8692fdb2e800 req-16fc8cbf-d525-4a91-afba-bb8a11fc841e service nova] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Refreshing instance network info cache due to event network-changed-5298015f-1096-4db7-8e70-1c91859e685f. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1167.570460] env[63021]: DEBUG oslo_concurrency.lockutils [req-894d1406-47a5-4752-a6b1-8692fdb2e800 req-16fc8cbf-d525-4a91-afba-bb8a11fc841e service nova] Acquiring lock "refresh_cache-e87e4421-d0a9-4d5d-816e-abba7caa687e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1167.570640] env[63021]: DEBUG oslo_concurrency.lockutils [req-894d1406-47a5-4752-a6b1-8692fdb2e800 req-16fc8cbf-d525-4a91-afba-bb8a11fc841e service nova] Acquired lock "refresh_cache-e87e4421-d0a9-4d5d-816e-abba7caa687e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1167.570856] env[63021]: DEBUG nova.network.neutron [req-894d1406-47a5-4752-a6b1-8692fdb2e800 req-16fc8cbf-d525-4a91-afba-bb8a11fc841e service nova] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Refreshing network info cache for port 5298015f-1096-4db7-8e70-1c91859e685f {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1167.913694] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294391, 'name': CreateVM_Task, 'duration_secs': 0.286667} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.914065] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1167.914672] env[63021]: DEBUG oslo_concurrency.lockutils [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1167.914926] env[63021]: DEBUG oslo_concurrency.lockutils [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1167.915348] env[63021]: DEBUG oslo_concurrency.lockutils [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1167.915651] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06ae8b47-a81a-4fa8-a3a6-7ed3acd5e2d9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.920214] env[63021]: DEBUG oslo_vmware.api [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1167.920214] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52b1a896-4dab-662b-290c-0ab07b36deb8" [ 1167.920214] env[63021]: _type = "Task" [ 1167.920214] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.928459] env[63021]: DEBUG oslo_vmware.api [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52b1a896-4dab-662b-290c-0ab07b36deb8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.256815] env[63021]: DEBUG nova.network.neutron [req-894d1406-47a5-4752-a6b1-8692fdb2e800 req-16fc8cbf-d525-4a91-afba-bb8a11fc841e service nova] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Updated VIF entry in instance network info cache for port 5298015f-1096-4db7-8e70-1c91859e685f. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1168.257253] env[63021]: DEBUG nova.network.neutron [req-894d1406-47a5-4752-a6b1-8692fdb2e800 req-16fc8cbf-d525-4a91-afba-bb8a11fc841e service nova] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Updating instance_info_cache with network_info: [{"id": "5298015f-1096-4db7-8e70-1c91859e685f", "address": "fa:16:3e:c8:03:c8", "network": {"id": "9502e364-60e1-47ef-a1f7-f3e3ff45cf69", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-164717784-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5af87713ae6b477ca8a584aed287889d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5298015f-10", "ovs_interfaceid": "5298015f-1096-4db7-8e70-1c91859e685f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1168.430665] env[63021]: DEBUG oslo_vmware.api [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52b1a896-4dab-662b-290c-0ab07b36deb8, 'name': SearchDatastore_Task, 'duration_secs': 0.011727} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.430936] env[63021]: DEBUG oslo_concurrency.lockutils [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1168.431215] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1168.431469] env[63021]: DEBUG oslo_concurrency.lockutils [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1168.431620] env[63021]: DEBUG oslo_concurrency.lockutils [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1168.431809] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1168.432089] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e4e3c92d-f7a5-4fd6-bbde-412de12aea41 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.456471] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1168.456643] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1168.457384] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14f2ea1d-ab47-4809-83cd-902825b43b17 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.462201] env[63021]: DEBUG oslo_vmware.api [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1168.462201] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52790d37-201f-de07-30d6-46e429c1d55b" [ 1168.462201] env[63021]: _type = "Task" [ 1168.462201] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.469196] env[63021]: DEBUG oslo_vmware.api [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52790d37-201f-de07-30d6-46e429c1d55b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.760486] env[63021]: DEBUG oslo_concurrency.lockutils [req-894d1406-47a5-4752-a6b1-8692fdb2e800 req-16fc8cbf-d525-4a91-afba-bb8a11fc841e service nova] Releasing lock "refresh_cache-e87e4421-d0a9-4d5d-816e-abba7caa687e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1168.972908] env[63021]: DEBUG oslo_vmware.api [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52790d37-201f-de07-30d6-46e429c1d55b, 'name': SearchDatastore_Task, 'duration_secs': 0.010392} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.973905] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4107b5e-9244-4069-9d26-b5702536e58f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.978752] env[63021]: DEBUG oslo_vmware.api [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1168.978752] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]5236e478-f69b-b8b4-df2a-ac649d923159" [ 1168.978752] env[63021]: _type = "Task" [ 1168.978752] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.986787] env[63021]: DEBUG oslo_vmware.api [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5236e478-f69b-b8b4-df2a-ac649d923159, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.488914] env[63021]: DEBUG oslo_vmware.api [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]5236e478-f69b-b8b4-df2a-ac649d923159, 'name': SearchDatastore_Task, 'duration_secs': 0.008672} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.489219] env[63021]: DEBUG oslo_concurrency.lockutils [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1169.489550] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] e87e4421-d0a9-4d5d-816e-abba7caa687e/e87e4421-d0a9-4d5d-816e-abba7caa687e.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1169.489824] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ea49d698-6670-4731-af65-89df873f9ada {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.496177] env[63021]: DEBUG oslo_vmware.api [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1169.496177] env[63021]: value = "task-1294392" [ 1169.496177] env[63021]: _type = "Task" [ 1169.496177] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.503527] env[63021]: DEBUG oslo_vmware.api [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294392, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.005824] env[63021]: DEBUG oslo_vmware.api [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294392, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.497403} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.006198] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] e87e4421-d0a9-4d5d-816e-abba7caa687e/e87e4421-d0a9-4d5d-816e-abba7caa687e.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1170.006356] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1170.006604] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b6a8400d-f232-4e9d-a4a3-7df091f9905a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.012257] env[63021]: DEBUG oslo_vmware.api [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1170.012257] env[63021]: value = "task-1294393" [ 1170.012257] env[63021]: _type = "Task" [ 1170.012257] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.019171] env[63021]: DEBUG oslo_vmware.api [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294393, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.522794] env[63021]: DEBUG oslo_vmware.api [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294393, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.05741} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.523055] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1170.523838] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-567f666b-d817-401c-830a-e9d47fea4808 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.547877] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] e87e4421-d0a9-4d5d-816e-abba7caa687e/e87e4421-d0a9-4d5d-816e-abba7caa687e.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1170.547877] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8595861a-a813-4c5a-b92e-4dd3c703ecb0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.566560] env[63021]: DEBUG oslo_vmware.api [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1170.566560] env[63021]: value = "task-1294394" [ 1170.566560] env[63021]: _type = "Task" [ 1170.566560] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.573795] env[63021]: DEBUG oslo_vmware.api [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294394, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.076800] env[63021]: DEBUG oslo_vmware.api [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294394, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.577219] env[63021]: DEBUG oslo_vmware.api [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294394, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.077502] env[63021]: DEBUG oslo_vmware.api [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294394, 'name': ReconfigVM_Task, 'duration_secs': 1.465915} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.077845] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Reconfigured VM instance instance-0000006e to attach disk [datastore1] e87e4421-d0a9-4d5d-816e-abba7caa687e/e87e4421-d0a9-4d5d-816e-abba7caa687e.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1172.078468] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-def13a9e-54b5-4e59-a1ec-e997d5fc3605 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.084324] env[63021]: DEBUG oslo_vmware.api [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1172.084324] env[63021]: value = "task-1294395" [ 1172.084324] env[63021]: _type = "Task" [ 1172.084324] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.091862] env[63021]: DEBUG oslo_vmware.api [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294395, 'name': Rename_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.593491] env[63021]: DEBUG oslo_vmware.api [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294395, 'name': Rename_Task, 'duration_secs': 0.131872} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.593725] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1172.593968] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5754c10d-a397-4635-b9c1-6837f2774b3a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.600369] env[63021]: DEBUG oslo_vmware.api [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1172.600369] env[63021]: value = "task-1294396" [ 1172.600369] env[63021]: _type = "Task" [ 1172.600369] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.607653] env[63021]: DEBUG oslo_vmware.api [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294396, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.111586] env[63021]: DEBUG oslo_vmware.api [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294396, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.611340] env[63021]: DEBUG oslo_vmware.api [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294396, 'name': PowerOnVM_Task, 'duration_secs': 0.841611} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.611608] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1173.611812] env[63021]: INFO nova.compute.manager [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Took 8.23 seconds to spawn the instance on the hypervisor. [ 1173.611993] env[63021]: DEBUG nova.compute.manager [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1173.612842] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de60c2f8-06c0-4fed-885a-d58915edc3b1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.131354] env[63021]: INFO nova.compute.manager [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Took 12.91 seconds to build instance. [ 1174.309773] env[63021]: DEBUG nova.compute.manager [req-84b4aa76-53a3-4482-a71c-3e57d8888146 req-d566a8ee-d5b0-43a1-a967-571de7c5ff19 service nova] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Received event network-changed-5298015f-1096-4db7-8e70-1c91859e685f {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1174.309977] env[63021]: DEBUG nova.compute.manager [req-84b4aa76-53a3-4482-a71c-3e57d8888146 req-d566a8ee-d5b0-43a1-a967-571de7c5ff19 service nova] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Refreshing instance network info cache due to event network-changed-5298015f-1096-4db7-8e70-1c91859e685f. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1174.310406] env[63021]: DEBUG oslo_concurrency.lockutils [req-84b4aa76-53a3-4482-a71c-3e57d8888146 req-d566a8ee-d5b0-43a1-a967-571de7c5ff19 service nova] Acquiring lock "refresh_cache-e87e4421-d0a9-4d5d-816e-abba7caa687e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1174.310575] env[63021]: DEBUG oslo_concurrency.lockutils [req-84b4aa76-53a3-4482-a71c-3e57d8888146 req-d566a8ee-d5b0-43a1-a967-571de7c5ff19 service nova] Acquired lock "refresh_cache-e87e4421-d0a9-4d5d-816e-abba7caa687e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1174.310740] env[63021]: DEBUG nova.network.neutron [req-84b4aa76-53a3-4482-a71c-3e57d8888146 req-d566a8ee-d5b0-43a1-a967-571de7c5ff19 service nova] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Refreshing network info cache for port 5298015f-1096-4db7-8e70-1c91859e685f {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1174.633832] env[63021]: DEBUG oslo_concurrency.lockutils [None req-55ccfe30-7e49-44cb-986a-157b1484f667 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "e87e4421-d0a9-4d5d-816e-abba7caa687e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.415s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1175.015393] env[63021]: DEBUG nova.network.neutron [req-84b4aa76-53a3-4482-a71c-3e57d8888146 req-d566a8ee-d5b0-43a1-a967-571de7c5ff19 service nova] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Updated VIF entry in instance network info cache for port 5298015f-1096-4db7-8e70-1c91859e685f. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1175.015760] env[63021]: DEBUG nova.network.neutron [req-84b4aa76-53a3-4482-a71c-3e57d8888146 req-d566a8ee-d5b0-43a1-a967-571de7c5ff19 service nova] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Updating instance_info_cache with network_info: [{"id": "5298015f-1096-4db7-8e70-1c91859e685f", "address": "fa:16:3e:c8:03:c8", "network": {"id": "9502e364-60e1-47ef-a1f7-f3e3ff45cf69", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-164717784-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5af87713ae6b477ca8a584aed287889d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5298015f-10", "ovs_interfaceid": "5298015f-1096-4db7-8e70-1c91859e685f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1175.518432] env[63021]: DEBUG oslo_concurrency.lockutils [req-84b4aa76-53a3-4482-a71c-3e57d8888146 req-d566a8ee-d5b0-43a1-a967-571de7c5ff19 service nova] Releasing lock "refresh_cache-e87e4421-d0a9-4d5d-816e-abba7caa687e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1204.667052] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bdd29811-5d98-421a-83a8-73a942025581 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "2e71d990-8bae-4ae8-b7c7-dd26984a08f5" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1204.667334] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bdd29811-5d98-421a-83a8-73a942025581 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "2e71d990-8bae-4ae8-b7c7-dd26984a08f5" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1205.170964] env[63021]: DEBUG nova.compute.utils [None req-bdd29811-5d98-421a-83a8-73a942025581 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1205.674383] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bdd29811-5d98-421a-83a8-73a942025581 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "2e71d990-8bae-4ae8-b7c7-dd26984a08f5" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1206.734087] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bdd29811-5d98-421a-83a8-73a942025581 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "2e71d990-8bae-4ae8-b7c7-dd26984a08f5" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1206.734454] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bdd29811-5d98-421a-83a8-73a942025581 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "2e71d990-8bae-4ae8-b7c7-dd26984a08f5" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1206.734650] env[63021]: INFO nova.compute.manager [None req-bdd29811-5d98-421a-83a8-73a942025581 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Attaching volume 26000887-4dde-4468-a0c3-1183f54bee4f to /dev/sdb [ 1206.764400] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5143b8ed-4880-4817-8d06-494599e5a93a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.773012] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bb1c5a1-37b2-4dd0-b2e0-468aab301940 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.785347] env[63021]: DEBUG nova.virt.block_device [None req-bdd29811-5d98-421a-83a8-73a942025581 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Updating existing volume attachment record: b8174010-9afc-4cad-a7ce-6629a427206f {{(pid=63021) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1211.327482] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-bdd29811-5d98-421a-83a8-73a942025581 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Volume attach. Driver type: vmdk {{(pid=63021) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1211.327782] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-bdd29811-5d98-421a-83a8-73a942025581 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277606', 'volume_id': '26000887-4dde-4468-a0c3-1183f54bee4f', 'name': 'volume-26000887-4dde-4468-a0c3-1183f54bee4f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2e71d990-8bae-4ae8-b7c7-dd26984a08f5', 'attached_at': '', 'detached_at': '', 'volume_id': '26000887-4dde-4468-a0c3-1183f54bee4f', 'serial': '26000887-4dde-4468-a0c3-1183f54bee4f'} {{(pid=63021) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1211.329112] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b54b137f-02d5-4fd0-aed9-353b37563dea {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.345735] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3bf0970-a18e-4421-8766-a4f1938305d6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.369454] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-bdd29811-5d98-421a-83a8-73a942025581 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] volume-26000887-4dde-4468-a0c3-1183f54bee4f/volume-26000887-4dde-4468-a0c3-1183f54bee4f.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1211.369726] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f6b30402-0739-457f-b17d-ecb4c041b842 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.387249] env[63021]: DEBUG oslo_vmware.api [None req-bdd29811-5d98-421a-83a8-73a942025581 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1211.387249] env[63021]: value = "task-1294399" [ 1211.387249] env[63021]: _type = "Task" [ 1211.387249] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.395086] env[63021]: DEBUG oslo_vmware.api [None req-bdd29811-5d98-421a-83a8-73a942025581 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294399, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.599978] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1211.600204] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63021) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1211.756150] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5793b667-0b52-4699-9c3a-fbde836e31f1 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "e87e4421-d0a9-4d5d-816e-abba7caa687e" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1211.756447] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5793b667-0b52-4699-9c3a-fbde836e31f1 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "e87e4421-d0a9-4d5d-816e-abba7caa687e" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1211.756634] env[63021]: DEBUG nova.compute.manager [None req-5793b667-0b52-4699-9c3a-fbde836e31f1 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1211.757534] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa1189ba-49c4-43c7-993d-1df077d17a69 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.764251] env[63021]: DEBUG nova.compute.manager [None req-5793b667-0b52-4699-9c3a-fbde836e31f1 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63021) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1211.764832] env[63021]: DEBUG nova.objects.instance [None req-5793b667-0b52-4699-9c3a-fbde836e31f1 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lazy-loading 'flavor' on Instance uuid e87e4421-d0a9-4d5d-816e-abba7caa687e {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1211.896138] env[63021]: DEBUG oslo_vmware.api [None req-bdd29811-5d98-421a-83a8-73a942025581 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294399, 'name': ReconfigVM_Task, 'duration_secs': 0.312477} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.896395] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-bdd29811-5d98-421a-83a8-73a942025581 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Reconfigured VM instance instance-0000006d to attach disk [datastore1] volume-26000887-4dde-4468-a0c3-1183f54bee4f/volume-26000887-4dde-4468-a0c3-1183f54bee4f.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1211.901173] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9c1dc08a-aa73-4bcf-90b9-bd2412d55d43 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.915842] env[63021]: DEBUG oslo_vmware.api [None req-bdd29811-5d98-421a-83a8-73a942025581 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1211.915842] env[63021]: value = "task-1294400" [ 1211.915842] env[63021]: _type = "Task" [ 1211.915842] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.923441] env[63021]: DEBUG oslo_vmware.api [None req-bdd29811-5d98-421a-83a8-73a942025581 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294400, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.269618] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-5793b667-0b52-4699-9c3a-fbde836e31f1 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1212.269891] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-056bc435-ff77-4502-a40a-e91c255df246 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.277560] env[63021]: DEBUG oslo_vmware.api [None req-5793b667-0b52-4699-9c3a-fbde836e31f1 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1212.277560] env[63021]: value = "task-1294401" [ 1212.277560] env[63021]: _type = "Task" [ 1212.277560] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.285281] env[63021]: DEBUG oslo_vmware.api [None req-5793b667-0b52-4699-9c3a-fbde836e31f1 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294401, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.425461] env[63021]: DEBUG oslo_vmware.api [None req-bdd29811-5d98-421a-83a8-73a942025581 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294400, 'name': ReconfigVM_Task, 'duration_secs': 0.149505} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.425852] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-bdd29811-5d98-421a-83a8-73a942025581 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277606', 'volume_id': '26000887-4dde-4468-a0c3-1183f54bee4f', 'name': 'volume-26000887-4dde-4468-a0c3-1183f54bee4f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2e71d990-8bae-4ae8-b7c7-dd26984a08f5', 'attached_at': '', 'detached_at': '', 'volume_id': '26000887-4dde-4468-a0c3-1183f54bee4f', 'serial': '26000887-4dde-4468-a0c3-1183f54bee4f'} {{(pid=63021) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1212.786731] env[63021]: DEBUG oslo_vmware.api [None req-5793b667-0b52-4699-9c3a-fbde836e31f1 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294401, 'name': PowerOffVM_Task, 'duration_secs': 0.190099} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.786995] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-5793b667-0b52-4699-9c3a-fbde836e31f1 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1212.787186] env[63021]: DEBUG nova.compute.manager [None req-5793b667-0b52-4699-9c3a-fbde836e31f1 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1212.787915] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ddc5dde-cd43-426a-a059-e84d6e5deaba {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.299133] env[63021]: DEBUG oslo_concurrency.lockutils [None req-5793b667-0b52-4699-9c3a-fbde836e31f1 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "e87e4421-d0a9-4d5d-816e-abba7caa687e" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.543s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1213.462969] env[63021]: DEBUG nova.objects.instance [None req-bdd29811-5d98-421a-83a8-73a942025581 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lazy-loading 'flavor' on Instance uuid 2e71d990-8bae-4ae8-b7c7-dd26984a08f5 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1213.595509] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1213.967755] env[63021]: DEBUG oslo_concurrency.lockutils [None req-bdd29811-5d98-421a-83a8-73a942025581 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "2e71d990-8bae-4ae8-b7c7-dd26984a08f5" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.233s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1214.182611] env[63021]: DEBUG oslo_concurrency.lockutils [None req-773e7b79-36f4-4609-9c57-f1b44ac47106 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "2e71d990-8bae-4ae8-b7c7-dd26984a08f5" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1214.182885] env[63021]: DEBUG oslo_concurrency.lockutils [None req-773e7b79-36f4-4609-9c57-f1b44ac47106 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "2e71d990-8bae-4ae8-b7c7-dd26984a08f5" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1214.391458] env[63021]: DEBUG nova.objects.instance [None req-2f636a72-0c95-4085-a206-2185bfcf85b2 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lazy-loading 'flavor' on Instance uuid e87e4421-d0a9-4d5d-816e-abba7caa687e {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1214.598913] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1214.599300] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1214.685716] env[63021]: INFO nova.compute.manager [None req-773e7b79-36f4-4609-9c57-f1b44ac47106 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Detaching volume 26000887-4dde-4468-a0c3-1183f54bee4f [ 1214.714771] env[63021]: INFO nova.virt.block_device [None req-773e7b79-36f4-4609-9c57-f1b44ac47106 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Attempting to driver detach volume 26000887-4dde-4468-a0c3-1183f54bee4f from mountpoint /dev/sdb [ 1214.715019] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-773e7b79-36f4-4609-9c57-f1b44ac47106 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Volume detach. Driver type: vmdk {{(pid=63021) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1214.715210] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-773e7b79-36f4-4609-9c57-f1b44ac47106 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277606', 'volume_id': '26000887-4dde-4468-a0c3-1183f54bee4f', 'name': 'volume-26000887-4dde-4468-a0c3-1183f54bee4f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2e71d990-8bae-4ae8-b7c7-dd26984a08f5', 'attached_at': '', 'detached_at': '', 'volume_id': '26000887-4dde-4468-a0c3-1183f54bee4f', 'serial': '26000887-4dde-4468-a0c3-1183f54bee4f'} {{(pid=63021) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1214.716107] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10d67988-a47e-4399-919d-539737c8b692 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.737820] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16c0d77e-32f0-45e7-bc9a-6788a17771e5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.744807] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22aff1f9-5670-4357-962e-e8839181028d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.763882] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19b1eafe-fbce-4e81-9bba-f1c3e44bafa1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.777497] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-773e7b79-36f4-4609-9c57-f1b44ac47106 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] The volume has not been displaced from its original location: [datastore1] volume-26000887-4dde-4468-a0c3-1183f54bee4f/volume-26000887-4dde-4468-a0c3-1183f54bee4f.vmdk. No consolidation needed. {{(pid=63021) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1214.782678] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-773e7b79-36f4-4609-9c57-f1b44ac47106 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Reconfiguring VM instance instance-0000006d to detach disk 2001 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1214.782935] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-35bbd5f4-4166-43a8-9fe8-aa506e7e03e9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.799708] env[63021]: DEBUG oslo_vmware.api [None req-773e7b79-36f4-4609-9c57-f1b44ac47106 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1214.799708] env[63021]: value = "task-1294402" [ 1214.799708] env[63021]: _type = "Task" [ 1214.799708] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.807053] env[63021]: DEBUG oslo_vmware.api [None req-773e7b79-36f4-4609-9c57-f1b44ac47106 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294402, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.896790] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2f636a72-0c95-4085-a206-2185bfcf85b2 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "refresh_cache-e87e4421-d0a9-4d5d-816e-abba7caa687e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1214.897028] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2f636a72-0c95-4085-a206-2185bfcf85b2 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquired lock "refresh_cache-e87e4421-d0a9-4d5d-816e-abba7caa687e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1214.897212] env[63021]: DEBUG nova.network.neutron [None req-2f636a72-0c95-4085-a206-2185bfcf85b2 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1214.897397] env[63021]: DEBUG nova.objects.instance [None req-2f636a72-0c95-4085-a206-2185bfcf85b2 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lazy-loading 'info_cache' on Instance uuid e87e4421-d0a9-4d5d-816e-abba7caa687e {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1215.102645] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1215.103021] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1215.103294] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1215.103541] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63021) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1215.105465] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12677e61-414a-483c-b3af-441826173cf6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.115279] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab5c037d-45c2-4f0b-9480-40d8babc6386 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.130354] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23ad0e41-33cd-43c2-b8b8-2f27e97a9b47 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.136263] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a33cac33-c5ee-4592-abfc-22add02351a4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.165355] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181462MB free_disk=138GB free_vcpus=48 pci_devices=None {{(pid=63021) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1215.165497] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1215.165682] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1215.309625] env[63021]: DEBUG oslo_vmware.api [None req-773e7b79-36f4-4609-9c57-f1b44ac47106 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294402, 'name': ReconfigVM_Task, 'duration_secs': 0.223943} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.309850] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-773e7b79-36f4-4609-9c57-f1b44ac47106 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Reconfigured VM instance instance-0000006d to detach disk 2001 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1215.314608] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6fbbcf96-899d-4d7e-93cd-5029f1731b90 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.329196] env[63021]: DEBUG oslo_vmware.api [None req-773e7b79-36f4-4609-9c57-f1b44ac47106 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1215.329196] env[63021]: value = "task-1294403" [ 1215.329196] env[63021]: _type = "Task" [ 1215.329196] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.336332] env[63021]: DEBUG oslo_vmware.api [None req-773e7b79-36f4-4609-9c57-f1b44ac47106 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294403, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.400573] env[63021]: DEBUG nova.objects.base [None req-2f636a72-0c95-4085-a206-2185bfcf85b2 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=63021) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1215.838678] env[63021]: DEBUG oslo_vmware.api [None req-773e7b79-36f4-4609-9c57-f1b44ac47106 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294403, 'name': ReconfigVM_Task, 'duration_secs': 0.137305} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.839073] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-773e7b79-36f4-4609-9c57-f1b44ac47106 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277606', 'volume_id': '26000887-4dde-4468-a0c3-1183f54bee4f', 'name': 'volume-26000887-4dde-4468-a0c3-1183f54bee4f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2e71d990-8bae-4ae8-b7c7-dd26984a08f5', 'attached_at': '', 'detached_at': '', 'volume_id': '26000887-4dde-4468-a0c3-1183f54bee4f', 'serial': '26000887-4dde-4468-a0c3-1183f54bee4f'} {{(pid=63021) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1216.101860] env[63021]: DEBUG nova.network.neutron [None req-2f636a72-0c95-4085-a206-2185bfcf85b2 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Updating instance_info_cache with network_info: [{"id": "5298015f-1096-4db7-8e70-1c91859e685f", "address": "fa:16:3e:c8:03:c8", "network": {"id": "9502e364-60e1-47ef-a1f7-f3e3ff45cf69", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-164717784-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5af87713ae6b477ca8a584aed287889d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5298015f-10", "ovs_interfaceid": "5298015f-1096-4db7-8e70-1c91859e685f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1216.193630] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 2e71d990-8bae-4ae8-b7c7-dd26984a08f5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1216.193816] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance e87e4421-d0a9-4d5d-816e-abba7caa687e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1216.194039] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=63021) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1216.194218] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=63021) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1216.228932] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b8e1b89-e06a-4f8c-bfeb-1042e7ca6dd0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.236296] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4eadd05-d2d8-46ea-8ef4-6b15f95e5b6a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.265860] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c82260e9-d91e-460a-bce7-49180f917282 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.272418] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7785bf68-76c5-4127-ac9b-5235405a3959 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.284687] env[63021]: DEBUG nova.compute.provider_tree [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Updating inventory in ProviderTree for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1216.378602] env[63021]: DEBUG nova.objects.instance [None req-773e7b79-36f4-4609-9c57-f1b44ac47106 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lazy-loading 'flavor' on Instance uuid 2e71d990-8bae-4ae8-b7c7-dd26984a08f5 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1216.605445] env[63021]: DEBUG oslo_concurrency.lockutils [None req-2f636a72-0c95-4085-a206-2185bfcf85b2 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Releasing lock "refresh_cache-e87e4421-d0a9-4d5d-816e-abba7caa687e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1216.803975] env[63021]: ERROR nova.scheduler.client.report [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [req-3bcb198c-4380-40f6-9b44-147bdf5643c0] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 6ef6c041-2b0d-4bc3-8662-5d679a77a86d. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-3bcb198c-4380-40f6-9b44-147bdf5643c0"}]} [ 1216.818942] env[63021]: DEBUG nova.scheduler.client.report [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Refreshing inventories for resource provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1216.831060] env[63021]: DEBUG nova.scheduler.client.report [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Updating ProviderTree inventory for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1216.831233] env[63021]: DEBUG nova.compute.provider_tree [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Updating inventory in ProviderTree for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1216.841101] env[63021]: DEBUG nova.scheduler.client.report [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Refreshing aggregate associations for resource provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d, aggregates: None {{(pid=63021) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1216.857554] env[63021]: DEBUG nova.scheduler.client.report [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Refreshing trait associations for resource provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE {{(pid=63021) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1216.888020] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edb66ade-f144-4708-8c5a-6d4bed8b2110 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.895279] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0b1ed13-61c7-465a-9618-5916980ec944 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.923872] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee56f2cc-678a-4b3d-9208-4911a609acf2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.930375] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ccc7ddb-208e-4c68-b817-7b05563da6a6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.944486] env[63021]: DEBUG nova.compute.provider_tree [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Updating inventory in ProviderTree for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1217.109019] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f636a72-0c95-4085-a206-2185bfcf85b2 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1217.109307] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-37f102f4-5355-4ed9-a6dc-235ae0df53b7 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.116123] env[63021]: DEBUG oslo_vmware.api [None req-2f636a72-0c95-4085-a206-2185bfcf85b2 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1217.116123] env[63021]: value = "task-1294404" [ 1217.116123] env[63021]: _type = "Task" [ 1217.116123] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1217.123242] env[63021]: DEBUG oslo_vmware.api [None req-2f636a72-0c95-4085-a206-2185bfcf85b2 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294404, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.385060] env[63021]: DEBUG oslo_concurrency.lockutils [None req-773e7b79-36f4-4609-9c57-f1b44ac47106 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "2e71d990-8bae-4ae8-b7c7-dd26984a08f5" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.202s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1217.477082] env[63021]: DEBUG nova.scheduler.client.report [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Updated inventory for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with generation 142 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1217.477411] env[63021]: DEBUG nova.compute.provider_tree [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Updating resource provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d generation from 142 to 143 during operation: update_inventory {{(pid=63021) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1217.477636] env[63021]: DEBUG nova.compute.provider_tree [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Updating inventory in ProviderTree for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1217.626992] env[63021]: DEBUG oslo_vmware.api [None req-2f636a72-0c95-4085-a206-2185bfcf85b2 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294404, 'name': PowerOnVM_Task, 'duration_secs': 0.359791} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1217.627286] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f636a72-0c95-4085-a206-2185bfcf85b2 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1217.627457] env[63021]: DEBUG nova.compute.manager [None req-2f636a72-0c95-4085-a206-2185bfcf85b2 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1217.628271] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e102be7e-3aba-4eeb-b609-f496a75d1abd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.983410] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63021) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1217.983779] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.818s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1218.419442] env[63021]: DEBUG oslo_concurrency.lockutils [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "2e71d990-8bae-4ae8-b7c7-dd26984a08f5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1218.419711] env[63021]: DEBUG oslo_concurrency.lockutils [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "2e71d990-8bae-4ae8-b7c7-dd26984a08f5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1218.419920] env[63021]: DEBUG oslo_concurrency.lockutils [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "2e71d990-8bae-4ae8-b7c7-dd26984a08f5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1218.420119] env[63021]: DEBUG oslo_concurrency.lockutils [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "2e71d990-8bae-4ae8-b7c7-dd26984a08f5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1218.420300] env[63021]: DEBUG oslo_concurrency.lockutils [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "2e71d990-8bae-4ae8-b7c7-dd26984a08f5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1218.422474] env[63021]: INFO nova.compute.manager [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Terminating instance [ 1218.424190] env[63021]: DEBUG nova.compute.manager [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1218.424391] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1218.425208] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa7a8311-9d32-44e6-b713-db08806ea7a1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.433082] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1218.433318] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f187b5f7-fdf2-47d0-9b42-fe18359f1bd2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.439814] env[63021]: DEBUG oslo_vmware.api [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1218.439814] env[63021]: value = "task-1294405" [ 1218.439814] env[63021]: _type = "Task" [ 1218.439814] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.447587] env[63021]: DEBUG oslo_vmware.api [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294405, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.951285] env[63021]: DEBUG oslo_vmware.api [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294405, 'name': PowerOffVM_Task, 'duration_secs': 0.21238} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1218.951595] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1218.951770] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1218.952034] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2348d73f-65f5-4c33-993c-a3f52874241b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.984869] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1219.775598] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1219.775598] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1219.775598] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Deleting the datastore file [datastore1] 2e71d990-8bae-4ae8-b7c7-dd26984a08f5 {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1219.775598] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8292dfc-d5e8-42c0-ab97-bbe6efd6eaec {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.777678] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1219.777835] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Starting heal instance info cache {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1219.777950] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Rebuilding the list of instances to heal {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1219.779386] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-748f4587-1f49-4117-b10a-5f9e0a3d42f6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.786044] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-de7e59f3-2388-4c53-9260-edaa8a0c37f2 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Suspending the VM {{(pid=63021) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1219.787121] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-1b7af697-d469-428e-8a91-1f4d5493126d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.788584] env[63021]: DEBUG oslo_vmware.api [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1219.788584] env[63021]: value = "task-1294407" [ 1219.788584] env[63021]: _type = "Task" [ 1219.788584] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1219.793495] env[63021]: DEBUG oslo_vmware.api [None req-de7e59f3-2388-4c53-9260-edaa8a0c37f2 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1219.793495] env[63021]: value = "task-1294408" [ 1219.793495] env[63021]: _type = "Task" [ 1219.793495] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1219.799231] env[63021]: DEBUG oslo_vmware.api [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294407, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.804150] env[63021]: DEBUG oslo_vmware.api [None req-de7e59f3-2388-4c53-9260-edaa8a0c37f2 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294408, 'name': SuspendVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.283025] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Skipping network cache update for instance because it is being deleted. {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 1220.283386] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquiring lock "refresh_cache-e87e4421-d0a9-4d5d-816e-abba7caa687e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1220.283435] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquired lock "refresh_cache-e87e4421-d0a9-4d5d-816e-abba7caa687e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1220.283557] env[63021]: DEBUG nova.network.neutron [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Forcefully refreshing network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1220.283705] env[63021]: DEBUG nova.objects.instance [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lazy-loading 'info_cache' on Instance uuid e87e4421-d0a9-4d5d-816e-abba7caa687e {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1220.297630] env[63021]: DEBUG oslo_vmware.api [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294407, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.23881} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.301361] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1220.301560] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1220.301735] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1220.301907] env[63021]: INFO nova.compute.manager [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Took 1.88 seconds to destroy the instance on the hypervisor. [ 1220.302155] env[63021]: DEBUG oslo.service.loopingcall [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1220.302594] env[63021]: DEBUG nova.compute.manager [-] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1220.302691] env[63021]: DEBUG nova.network.neutron [-] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1220.308551] env[63021]: DEBUG oslo_vmware.api [None req-de7e59f3-2388-4c53-9260-edaa8a0c37f2 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294408, 'name': SuspendVM_Task} progress is 58%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.760637] env[63021]: DEBUG nova.compute.manager [req-7eb556fd-2cf7-40f2-a54d-0b48fd0ffb73 req-6c2a79e2-f126-433d-8fc1-63261135484e service nova] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Received event network-vif-deleted-e9488dcd-9157-488c-8f45-8c497d12ea22 {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1220.760848] env[63021]: INFO nova.compute.manager [req-7eb556fd-2cf7-40f2-a54d-0b48fd0ffb73 req-6c2a79e2-f126-433d-8fc1-63261135484e service nova] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Neutron deleted interface e9488dcd-9157-488c-8f45-8c497d12ea22; detaching it from the instance and deleting it from the info cache [ 1220.761032] env[63021]: DEBUG nova.network.neutron [req-7eb556fd-2cf7-40f2-a54d-0b48fd0ffb73 req-6c2a79e2-f126-433d-8fc1-63261135484e service nova] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1220.806806] env[63021]: DEBUG oslo_vmware.api [None req-de7e59f3-2388-4c53-9260-edaa8a0c37f2 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294408, 'name': SuspendVM_Task, 'duration_secs': 0.622593} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.807077] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-de7e59f3-2388-4c53-9260-edaa8a0c37f2 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Suspended the VM {{(pid=63021) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1220.807260] env[63021]: DEBUG nova.compute.manager [None req-de7e59f3-2388-4c53-9260-edaa8a0c37f2 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1220.808016] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-175eac42-be1e-42d0-bae6-ab165fa76d05 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.235215] env[63021]: DEBUG nova.network.neutron [-] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1221.264477] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eecbf58b-5336-46c9-be48-011bba70a17f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.274128] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eaf6c58-ea36-4686-b879-eaaf62a6d0ae {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.296958] env[63021]: DEBUG nova.compute.manager [req-7eb556fd-2cf7-40f2-a54d-0b48fd0ffb73 req-6c2a79e2-f126-433d-8fc1-63261135484e service nova] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Detach interface failed, port_id=e9488dcd-9157-488c-8f45-8c497d12ea22, reason: Instance 2e71d990-8bae-4ae8-b7c7-dd26984a08f5 could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1221.682216] env[63021]: INFO nova.compute.manager [None req-9dc400d4-d8c4-4e18-848c-e4b5e20b7fe9 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Resuming [ 1221.682847] env[63021]: DEBUG nova.objects.instance [None req-9dc400d4-d8c4-4e18-848c-e4b5e20b7fe9 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lazy-loading 'flavor' on Instance uuid e87e4421-d0a9-4d5d-816e-abba7caa687e {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1221.737892] env[63021]: INFO nova.compute.manager [-] [instance: 2e71d990-8bae-4ae8-b7c7-dd26984a08f5] Took 1.44 seconds to deallocate network for instance. [ 1221.994448] env[63021]: DEBUG nova.network.neutron [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Updating instance_info_cache with network_info: [{"id": "5298015f-1096-4db7-8e70-1c91859e685f", "address": "fa:16:3e:c8:03:c8", "network": {"id": "9502e364-60e1-47ef-a1f7-f3e3ff45cf69", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-164717784-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5af87713ae6b477ca8a584aed287889d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5298015f-10", "ovs_interfaceid": "5298015f-1096-4db7-8e70-1c91859e685f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1222.243654] env[63021]: DEBUG oslo_concurrency.lockutils [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1222.243943] env[63021]: DEBUG oslo_concurrency.lockutils [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1222.244150] env[63021]: DEBUG nova.objects.instance [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lazy-loading 'resources' on Instance uuid 2e71d990-8bae-4ae8-b7c7-dd26984a08f5 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1222.498112] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Releasing lock "refresh_cache-e87e4421-d0a9-4d5d-816e-abba7caa687e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1222.498112] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Updated the network info_cache for instance {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1222.498112] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1222.498112] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1222.498610] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1222.498610] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1222.690342] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9dc400d4-d8c4-4e18-848c-e4b5e20b7fe9 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "refresh_cache-e87e4421-d0a9-4d5d-816e-abba7caa687e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1222.690525] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9dc400d4-d8c4-4e18-848c-e4b5e20b7fe9 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquired lock "refresh_cache-e87e4421-d0a9-4d5d-816e-abba7caa687e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1222.690701] env[63021]: DEBUG nova.network.neutron [None req-9dc400d4-d8c4-4e18-848c-e4b5e20b7fe9 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1222.787169] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a32d64f8-2368-4b15-add3-1d73d4819ae3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.794674] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75c080f8-4715-4b2d-9a0e-85bfcfb50841 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.823570] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80636e03-d09a-4ad3-948b-510c797ca75d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.829853] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-201cf268-3a35-43e8-9384-b9ed04fba205 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.842393] env[63021]: DEBUG nova.compute.provider_tree [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1223.345648] env[63021]: DEBUG nova.scheduler.client.report [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1223.381828] env[63021]: DEBUG nova.network.neutron [None req-9dc400d4-d8c4-4e18-848c-e4b5e20b7fe9 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Updating instance_info_cache with network_info: [{"id": "5298015f-1096-4db7-8e70-1c91859e685f", "address": "fa:16:3e:c8:03:c8", "network": {"id": "9502e364-60e1-47ef-a1f7-f3e3ff45cf69", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-164717784-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5af87713ae6b477ca8a584aed287889d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5298015f-10", "ovs_interfaceid": "5298015f-1096-4db7-8e70-1c91859e685f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1223.852437] env[63021]: DEBUG oslo_concurrency.lockutils [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.608s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1223.879610] env[63021]: INFO nova.scheduler.client.report [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Deleted allocations for instance 2e71d990-8bae-4ae8-b7c7-dd26984a08f5 [ 1223.884566] env[63021]: DEBUG oslo_concurrency.lockutils [None req-9dc400d4-d8c4-4e18-848c-e4b5e20b7fe9 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Releasing lock "refresh_cache-e87e4421-d0a9-4d5d-816e-abba7caa687e" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1223.885456] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37d35bac-e735-442d-99cc-7c66958901d2 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.892108] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-9dc400d4-d8c4-4e18-848c-e4b5e20b7fe9 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Resuming the VM {{(pid=63021) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1223.892335] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-83f5be7b-f1c5-43bf-ad05-f67813e48125 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.898693] env[63021]: DEBUG oslo_vmware.api [None req-9dc400d4-d8c4-4e18-848c-e4b5e20b7fe9 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1223.898693] env[63021]: value = "task-1294409" [ 1223.898693] env[63021]: _type = "Task" [ 1223.898693] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.905989] env[63021]: DEBUG oslo_vmware.api [None req-9dc400d4-d8c4-4e18-848c-e4b5e20b7fe9 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294409, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.387759] env[63021]: DEBUG oslo_concurrency.lockutils [None req-573b50c8-9e3c-4b4e-802b-43e5e2be42f4 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "2e71d990-8bae-4ae8-b7c7-dd26984a08f5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.968s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1224.412360] env[63021]: DEBUG oslo_vmware.api [None req-9dc400d4-d8c4-4e18-848c-e4b5e20b7fe9 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294409, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.916090] env[63021]: DEBUG oslo_vmware.api [None req-9dc400d4-d8c4-4e18-848c-e4b5e20b7fe9 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294409, 'name': PowerOnVM_Task, 'duration_secs': 0.530199} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.916090] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-9dc400d4-d8c4-4e18-848c-e4b5e20b7fe9 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Resumed the VM {{(pid=63021) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1224.916090] env[63021]: DEBUG nova.compute.manager [None req-9dc400d4-d8c4-4e18-848c-e4b5e20b7fe9 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1224.916090] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b48e9e1-648d-4542-bc9c-6a22034bb02b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.633720] env[63021]: DEBUG oslo_concurrency.lockutils [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "5ee0e2b0-289d-47aa-b1a9-0db809ebbab4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1226.633996] env[63021]: DEBUG oslo_concurrency.lockutils [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "5ee0e2b0-289d-47aa-b1a9-0db809ebbab4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1226.673407] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "e87e4421-d0a9-4d5d-816e-abba7caa687e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1226.673695] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "e87e4421-d0a9-4d5d-816e-abba7caa687e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1226.673935] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "e87e4421-d0a9-4d5d-816e-abba7caa687e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1226.674172] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "e87e4421-d0a9-4d5d-816e-abba7caa687e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1226.674410] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "e87e4421-d0a9-4d5d-816e-abba7caa687e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1226.676528] env[63021]: INFO nova.compute.manager [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Terminating instance [ 1226.678441] env[63021]: DEBUG nova.compute.manager [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1226.678664] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1226.679621] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7617122-44a1-4107-b5e1-16fd82246226 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.687425] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1226.687648] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-34a2812d-7153-413c-beee-84e7e5bda011 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.693772] env[63021]: DEBUG oslo_vmware.api [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1226.693772] env[63021]: value = "task-1294410" [ 1226.693772] env[63021]: _type = "Task" [ 1226.693772] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.701155] env[63021]: DEBUG oslo_vmware.api [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294410, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.136657] env[63021]: DEBUG nova.compute.manager [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Starting instance... {{(pid=63021) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1227.203289] env[63021]: DEBUG oslo_vmware.api [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294410, 'name': PowerOffVM_Task, 'duration_secs': 0.214753} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.203591] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1227.203764] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1227.204020] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3556f8b9-41ca-469d-9de9-653e220f882d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.269530] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1227.269754] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1227.269938] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Deleting the datastore file [datastore1] e87e4421-d0a9-4d5d-816e-abba7caa687e {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1227.270223] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-992c69ea-5e22-4def-ac58-3172058ab467 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.276098] env[63021]: DEBUG oslo_vmware.api [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for the task: (returnval){ [ 1227.276098] env[63021]: value = "task-1294412" [ 1227.276098] env[63021]: _type = "Task" [ 1227.276098] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.283692] env[63021]: DEBUG oslo_vmware.api [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294412, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.658919] env[63021]: DEBUG oslo_concurrency.lockutils [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1227.659202] env[63021]: DEBUG oslo_concurrency.lockutils [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1227.660784] env[63021]: INFO nova.compute.claims [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1227.785197] env[63021]: DEBUG oslo_vmware.api [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Task: {'id': task-1294412, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.211085} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.785454] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1227.785644] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1227.785821] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1227.785992] env[63021]: INFO nova.compute.manager [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1227.786251] env[63021]: DEBUG oslo.service.loopingcall [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1227.786465] env[63021]: DEBUG nova.compute.manager [-] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1227.786533] env[63021]: DEBUG nova.network.neutron [-] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1228.229799] env[63021]: DEBUG nova.compute.manager [req-5d8a621e-7d42-4a9f-8447-96d9f2d3bf2f req-63218535-195d-4a89-afa3-b7ce1b056f17 service nova] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Received event network-vif-deleted-5298015f-1096-4db7-8e70-1c91859e685f {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1228.229989] env[63021]: INFO nova.compute.manager [req-5d8a621e-7d42-4a9f-8447-96d9f2d3bf2f req-63218535-195d-4a89-afa3-b7ce1b056f17 service nova] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Neutron deleted interface 5298015f-1096-4db7-8e70-1c91859e685f; detaching it from the instance and deleting it from the info cache [ 1228.230171] env[63021]: DEBUG nova.network.neutron [req-5d8a621e-7d42-4a9f-8447-96d9f2d3bf2f req-63218535-195d-4a89-afa3-b7ce1b056f17 service nova] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1228.708461] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b04382d2-f261-47d1-8c1a-aa05e904877a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.711083] env[63021]: DEBUG nova.network.neutron [-] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1228.716885] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ff2a492-9c7e-4564-af27-e2e6e3ce4205 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.747075] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-90314d11-8a51-4f44-9722-0efad6bee1dd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.749723] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ffd34ee-260a-4c59-8459-24ea1131f8e9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.758182] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c75dfdce-3198-4c38-b748-278aef16b6c1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.764491] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7ec7d60-f26e-4879-ac6d-6a2c2593cada {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.785575] env[63021]: DEBUG nova.compute.provider_tree [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Updating inventory in ProviderTree for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1228.790067] env[63021]: DEBUG nova.compute.manager [req-5d8a621e-7d42-4a9f-8447-96d9f2d3bf2f req-63218535-195d-4a89-afa3-b7ce1b056f17 service nova] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Detach interface failed, port_id=5298015f-1096-4db7-8e70-1c91859e685f, reason: Instance e87e4421-d0a9-4d5d-816e-abba7caa687e could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1229.214792] env[63021]: INFO nova.compute.manager [-] [instance: e87e4421-d0a9-4d5d-816e-abba7caa687e] Took 1.43 seconds to deallocate network for instance. [ 1229.308522] env[63021]: ERROR nova.scheduler.client.report [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [req-de6446a8-0346-4ed4-a927-9af0d9ac8987] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 6ef6c041-2b0d-4bc3-8662-5d679a77a86d. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-de6446a8-0346-4ed4-a927-9af0d9ac8987"}]} [ 1229.323914] env[63021]: DEBUG nova.scheduler.client.report [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Refreshing inventories for resource provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1229.336093] env[63021]: DEBUG nova.scheduler.client.report [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Updating ProviderTree inventory for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1229.336305] env[63021]: DEBUG nova.compute.provider_tree [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Updating inventory in ProviderTree for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1229.345930] env[63021]: DEBUG nova.scheduler.client.report [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Refreshing aggregate associations for resource provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d, aggregates: None {{(pid=63021) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1229.361991] env[63021]: DEBUG nova.scheduler.client.report [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Refreshing trait associations for resource provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE {{(pid=63021) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1229.392957] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55bef757-15c2-4a26-8f53-6fd61929eccc {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.400135] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d817efc2-7dc3-4368-8ed2-0be9cfc697bd {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.429227] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acc396c5-9488-49fc-a1c1-bc29f4a8c4b3 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.436363] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-630a161e-7f6d-487b-a91d-eecb644ceb6a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.449465] env[63021]: DEBUG nova.compute.provider_tree [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Updating inventory in ProviderTree for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1229.721893] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1229.980008] env[63021]: DEBUG nova.scheduler.client.report [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Updated inventory for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with generation 144 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1229.980318] env[63021]: DEBUG nova.compute.provider_tree [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Updating resource provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d generation from 144 to 145 during operation: update_inventory {{(pid=63021) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1229.980503] env[63021]: DEBUG nova.compute.provider_tree [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Updating inventory in ProviderTree for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1230.485390] env[63021]: DEBUG oslo_concurrency.lockutils [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.826s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1230.486009] env[63021]: DEBUG nova.compute.manager [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Start building networks asynchronously for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1230.488995] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.767s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1230.489255] env[63021]: DEBUG nova.objects.instance [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lazy-loading 'resources' on Instance uuid e87e4421-d0a9-4d5d-816e-abba7caa687e {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1230.992025] env[63021]: DEBUG nova.compute.utils [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1230.993543] env[63021]: DEBUG nova.compute.manager [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Allocating IP information in the background. {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1230.993714] env[63021]: DEBUG nova.network.neutron [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] allocate_for_instance() {{(pid=63021) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1231.031893] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-652c505f-98d0-40bf-b0dd-b559b36e2ec0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.039806] env[63021]: DEBUG nova.policy [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5431727705ff4c5483cd9a5b9cd104f6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '547d9a3099b94ff488ef6929ae591bac', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63021) authorize /opt/stack/nova/nova/policy.py:201}} [ 1231.041830] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68b8ab2f-7569-4e1d-929e-09ba4d46c65a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.071153] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffe08d79-3b6b-429f-a5c6-af05bdc6bab6 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.079606] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75719532-a7d1-46b9-894d-3c25eea303d0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.092205] env[63021]: DEBUG nova.compute.provider_tree [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1231.300438] env[63021]: DEBUG nova.network.neutron [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Successfully created port: 22b596bf-eee5-4b76-b9e8-ae3dd028c5ed {{(pid=63021) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1231.500742] env[63021]: DEBUG nova.compute.manager [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Start building block device mappings for instance. {{(pid=63021) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1231.594919] env[63021]: DEBUG nova.scheduler.client.report [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1232.099742] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.611s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1232.122815] env[63021]: INFO nova.scheduler.client.report [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Deleted allocations for instance e87e4421-d0a9-4d5d-816e-abba7caa687e [ 1232.511247] env[63021]: DEBUG nova.compute.manager [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Start spawning the instance on the hypervisor. {{(pid=63021) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1232.535060] env[63021]: DEBUG nova.virt.hardware [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T17:42:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T17:42:05Z,direct_url=,disk_format='vmdk',id=e624225d-af0f-415c-90cc-d4b228c479f6,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cc50c08b679248798f76b07c1530ed7e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T17:42:06Z,virtual_size=,visibility=), allow threads: False {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1232.535356] env[63021]: DEBUG nova.virt.hardware [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Flavor limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1232.535528] env[63021]: DEBUG nova.virt.hardware [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Image limits 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1232.535730] env[63021]: DEBUG nova.virt.hardware [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Flavor pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1232.535893] env[63021]: DEBUG nova.virt.hardware [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Image pref 0:0:0 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1232.536038] env[63021]: DEBUG nova.virt.hardware [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63021) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1232.536250] env[63021]: DEBUG nova.virt.hardware [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1232.536412] env[63021]: DEBUG nova.virt.hardware [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1232.536580] env[63021]: DEBUG nova.virt.hardware [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Got 1 possible topologies {{(pid=63021) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1232.536762] env[63021]: DEBUG nova.virt.hardware [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1232.536909] env[63021]: DEBUG nova.virt.hardware [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63021) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1232.537763] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3dca1de-fc0a-4f84-884a-55e63d44a824 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.545800] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56b42759-fd09-4a41-9db7-73996d621808 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.630672] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ea961fcb-1f8b-406f-8617-13997bba46a0 tempest-ServerActionsTestJSON-1621975545 tempest-ServerActionsTestJSON-1621975545-project-member] Lock "e87e4421-d0a9-4d5d-816e-abba7caa687e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.957s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1232.667119] env[63021]: DEBUG nova.compute.manager [req-8940bff1-629e-4a54-bb72-481a2ebd3e8a req-4c67b442-57b8-4529-b24a-23d2a2f7f430 service nova] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Received event network-vif-plugged-22b596bf-eee5-4b76-b9e8-ae3dd028c5ed {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1232.667285] env[63021]: DEBUG oslo_concurrency.lockutils [req-8940bff1-629e-4a54-bb72-481a2ebd3e8a req-4c67b442-57b8-4529-b24a-23d2a2f7f430 service nova] Acquiring lock "5ee0e2b0-289d-47aa-b1a9-0db809ebbab4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1232.667496] env[63021]: DEBUG oslo_concurrency.lockutils [req-8940bff1-629e-4a54-bb72-481a2ebd3e8a req-4c67b442-57b8-4529-b24a-23d2a2f7f430 service nova] Lock "5ee0e2b0-289d-47aa-b1a9-0db809ebbab4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1232.667667] env[63021]: DEBUG oslo_concurrency.lockutils [req-8940bff1-629e-4a54-bb72-481a2ebd3e8a req-4c67b442-57b8-4529-b24a-23d2a2f7f430 service nova] Lock "5ee0e2b0-289d-47aa-b1a9-0db809ebbab4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1232.667869] env[63021]: DEBUG nova.compute.manager [req-8940bff1-629e-4a54-bb72-481a2ebd3e8a req-4c67b442-57b8-4529-b24a-23d2a2f7f430 service nova] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] No waiting events found dispatching network-vif-plugged-22b596bf-eee5-4b76-b9e8-ae3dd028c5ed {{(pid=63021) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1232.667996] env[63021]: WARNING nova.compute.manager [req-8940bff1-629e-4a54-bb72-481a2ebd3e8a req-4c67b442-57b8-4529-b24a-23d2a2f7f430 service nova] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Received unexpected event network-vif-plugged-22b596bf-eee5-4b76-b9e8-ae3dd028c5ed for instance with vm_state building and task_state spawning. [ 1232.759149] env[63021]: DEBUG nova.network.neutron [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Successfully updated port: 22b596bf-eee5-4b76-b9e8-ae3dd028c5ed {{(pid=63021) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1233.263334] env[63021]: DEBUG oslo_concurrency.lockutils [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "refresh_cache-5ee0e2b0-289d-47aa-b1a9-0db809ebbab4" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1233.263334] env[63021]: DEBUG oslo_concurrency.lockutils [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquired lock "refresh_cache-5ee0e2b0-289d-47aa-b1a9-0db809ebbab4" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1233.263334] env[63021]: DEBUG nova.network.neutron [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Building network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1233.792164] env[63021]: DEBUG nova.network.neutron [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Instance cache missing network info. {{(pid=63021) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1233.914428] env[63021]: DEBUG nova.network.neutron [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Updating instance_info_cache with network_info: [{"id": "22b596bf-eee5-4b76-b9e8-ae3dd028c5ed", "address": "fa:16:3e:a5:25:a9", "network": {"id": "d27bb3cc-fb80-4b03-9288-ed17a0d05140", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-122910456-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "547d9a3099b94ff488ef6929ae591bac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1c797172-a569-458e-aeb0-3f21e589a740", "external-id": "nsx-vlan-transportzone-957", "segmentation_id": 957, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22b596bf-ee", "ovs_interfaceid": "22b596bf-eee5-4b76-b9e8-ae3dd028c5ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1234.416836] env[63021]: DEBUG oslo_concurrency.lockutils [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Releasing lock "refresh_cache-5ee0e2b0-289d-47aa-b1a9-0db809ebbab4" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1234.417132] env[63021]: DEBUG nova.compute.manager [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Instance network_info: |[{"id": "22b596bf-eee5-4b76-b9e8-ae3dd028c5ed", "address": "fa:16:3e:a5:25:a9", "network": {"id": "d27bb3cc-fb80-4b03-9288-ed17a0d05140", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-122910456-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "547d9a3099b94ff488ef6929ae591bac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1c797172-a569-458e-aeb0-3f21e589a740", "external-id": "nsx-vlan-transportzone-957", "segmentation_id": 957, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22b596bf-ee", "ovs_interfaceid": "22b596bf-eee5-4b76-b9e8-ae3dd028c5ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63021) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1234.418795] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a5:25:a9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1c797172-a569-458e-aeb0-3f21e589a740', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '22b596bf-eee5-4b76-b9e8-ae3dd028c5ed', 'vif_model': 'vmxnet3'}] {{(pid=63021) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1234.424961] env[63021]: DEBUG oslo.service.loopingcall [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1234.425201] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Creating VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1234.425426] env[63021]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f6aa071e-d942-4361-9a11-472634e7ed0f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.445933] env[63021]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1234.445933] env[63021]: value = "task-1294413" [ 1234.445933] env[63021]: _type = "Task" [ 1234.445933] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.453352] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294413, 'name': CreateVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.699532] env[63021]: DEBUG nova.compute.manager [req-b9bc57ea-4bf4-4614-89ef-1cede709bf04 req-12f15710-0ad5-4179-95ef-d5570f0f5466 service nova] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Received event network-changed-22b596bf-eee5-4b76-b9e8-ae3dd028c5ed {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1234.699787] env[63021]: DEBUG nova.compute.manager [req-b9bc57ea-4bf4-4614-89ef-1cede709bf04 req-12f15710-0ad5-4179-95ef-d5570f0f5466 service nova] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Refreshing instance network info cache due to event network-changed-22b596bf-eee5-4b76-b9e8-ae3dd028c5ed. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1234.700068] env[63021]: DEBUG oslo_concurrency.lockutils [req-b9bc57ea-4bf4-4614-89ef-1cede709bf04 req-12f15710-0ad5-4179-95ef-d5570f0f5466 service nova] Acquiring lock "refresh_cache-5ee0e2b0-289d-47aa-b1a9-0db809ebbab4" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1234.700266] env[63021]: DEBUG oslo_concurrency.lockutils [req-b9bc57ea-4bf4-4614-89ef-1cede709bf04 req-12f15710-0ad5-4179-95ef-d5570f0f5466 service nova] Acquired lock "refresh_cache-5ee0e2b0-289d-47aa-b1a9-0db809ebbab4" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1234.700474] env[63021]: DEBUG nova.network.neutron [req-b9bc57ea-4bf4-4614-89ef-1cede709bf04 req-12f15710-0ad5-4179-95ef-d5570f0f5466 service nova] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Refreshing network info cache for port 22b596bf-eee5-4b76-b9e8-ae3dd028c5ed {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1234.956510] env[63021]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294413, 'name': CreateVM_Task, 'duration_secs': 0.306738} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1234.956700] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Created VM on the ESX host {{(pid=63021) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1234.957326] env[63021]: DEBUG oslo_concurrency.lockutils [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1234.957500] env[63021]: DEBUG oslo_concurrency.lockutils [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1234.957834] env[63021]: DEBUG oslo_concurrency.lockutils [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1234.958106] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4b57751-eaf6-4710-895c-b38a3f8f07de {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.962333] env[63021]: DEBUG oslo_vmware.api [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1234.962333] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]524f039a-0b10-73ca-d8e2-be08db1059d1" [ 1234.962333] env[63021]: _type = "Task" [ 1234.962333] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.970751] env[63021]: DEBUG oslo_vmware.api [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]524f039a-0b10-73ca-d8e2-be08db1059d1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.383566] env[63021]: DEBUG nova.network.neutron [req-b9bc57ea-4bf4-4614-89ef-1cede709bf04 req-12f15710-0ad5-4179-95ef-d5570f0f5466 service nova] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Updated VIF entry in instance network info cache for port 22b596bf-eee5-4b76-b9e8-ae3dd028c5ed. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1235.383923] env[63021]: DEBUG nova.network.neutron [req-b9bc57ea-4bf4-4614-89ef-1cede709bf04 req-12f15710-0ad5-4179-95ef-d5570f0f5466 service nova] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Updating instance_info_cache with network_info: [{"id": "22b596bf-eee5-4b76-b9e8-ae3dd028c5ed", "address": "fa:16:3e:a5:25:a9", "network": {"id": "d27bb3cc-fb80-4b03-9288-ed17a0d05140", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-122910456-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "547d9a3099b94ff488ef6929ae591bac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1c797172-a569-458e-aeb0-3f21e589a740", "external-id": "nsx-vlan-transportzone-957", "segmentation_id": 957, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22b596bf-ee", "ovs_interfaceid": "22b596bf-eee5-4b76-b9e8-ae3dd028c5ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1235.472645] env[63021]: DEBUG oslo_vmware.api [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]524f039a-0b10-73ca-d8e2-be08db1059d1, 'name': SearchDatastore_Task, 'duration_secs': 0.014314} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.472985] env[63021]: DEBUG oslo_concurrency.lockutils [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1235.473175] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Processing image e624225d-af0f-415c-90cc-d4b228c479f6 {{(pid=63021) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1235.473407] env[63021]: DEBUG oslo_concurrency.lockutils [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1235.473558] env[63021]: DEBUG oslo_concurrency.lockutils [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1235.473748] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1235.473999] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-641a6fa6-6a9d-4842-afd8-013f1b41d421 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.481848] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63021) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1235.481996] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63021) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1235.482668] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31c183a3-b83a-4c7e-9736-5a083e3785be {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.487336] env[63021]: DEBUG oslo_vmware.api [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1235.487336] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]526ba13d-e5d2-6afe-f7ba-4dbfec87665a" [ 1235.487336] env[63021]: _type = "Task" [ 1235.487336] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.494113] env[63021]: DEBUG oslo_vmware.api [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]526ba13d-e5d2-6afe-f7ba-4dbfec87665a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.886597] env[63021]: DEBUG oslo_concurrency.lockutils [req-b9bc57ea-4bf4-4614-89ef-1cede709bf04 req-12f15710-0ad5-4179-95ef-d5570f0f5466 service nova] Releasing lock "refresh_cache-5ee0e2b0-289d-47aa-b1a9-0db809ebbab4" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1235.998100] env[63021]: DEBUG oslo_vmware.api [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]526ba13d-e5d2-6afe-f7ba-4dbfec87665a, 'name': SearchDatastore_Task, 'duration_secs': 0.009156} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.998899] env[63021]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73a4db61-6754-4081-880a-d3d4aa1e6682 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.003859] env[63021]: DEBUG oslo_vmware.api [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1236.003859] env[63021]: value = "session[529561b8-a441-144b-8d80-b21567909a83]52a6c15d-4419-0d7d-fda8-51ca9f18f43c" [ 1236.003859] env[63021]: _type = "Task" [ 1236.003859] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.010633] env[63021]: DEBUG oslo_vmware.api [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52a6c15d-4419-0d7d-fda8-51ca9f18f43c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.513980] env[63021]: DEBUG oslo_vmware.api [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': session[529561b8-a441-144b-8d80-b21567909a83]52a6c15d-4419-0d7d-fda8-51ca9f18f43c, 'name': SearchDatastore_Task, 'duration_secs': 0.011283} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.514338] env[63021]: DEBUG oslo_concurrency.lockutils [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1236.514517] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4/5ee0e2b0-289d-47aa-b1a9-0db809ebbab4.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1236.514775] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1a04beca-7043-4a7c-95c8-ae1b9a96f54c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.520887] env[63021]: DEBUG oslo_vmware.api [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1236.520887] env[63021]: value = "task-1294414" [ 1236.520887] env[63021]: _type = "Task" [ 1236.520887] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.527956] env[63021]: DEBUG oslo_vmware.api [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294414, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1237.030639] env[63021]: DEBUG oslo_vmware.api [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294414, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1237.531944] env[63021]: DEBUG oslo_vmware.api [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294414, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.033058] env[63021]: DEBUG oslo_vmware.api [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294414, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.533831] env[63021]: DEBUG oslo_vmware.api [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294414, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.526162} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1238.534229] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e624225d-af0f-415c-90cc-d4b228c479f6/e624225d-af0f-415c-90cc-d4b228c479f6.vmdk to [datastore1] 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4/5ee0e2b0-289d-47aa-b1a9-0db809ebbab4.vmdk {{(pid=63021) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1238.534426] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Extending root virtual disk to 1048576 {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1238.534706] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a7524cd5-226c-4625-b6f6-c3e786d98a24 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.541759] env[63021]: DEBUG oslo_vmware.api [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1238.541759] env[63021]: value = "task-1294415" [ 1238.541759] env[63021]: _type = "Task" [ 1238.541759] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.549159] env[63021]: DEBUG oslo_vmware.api [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294415, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.052222] env[63021]: DEBUG oslo_vmware.api [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294415, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067113} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.052541] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Extended root virtual disk {{(pid=63021) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1239.053397] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8df0a68-aadd-4438-b027-f9855778a396 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.076708] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4/5ee0e2b0-289d-47aa-b1a9-0db809ebbab4.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1239.076968] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-72b87b80-5538-446a-a640-ed5056746db4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.096487] env[63021]: DEBUG oslo_vmware.api [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1239.096487] env[63021]: value = "task-1294416" [ 1239.096487] env[63021]: _type = "Task" [ 1239.096487] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.103514] env[63021]: DEBUG oslo_vmware.api [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294416, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.606699] env[63021]: DEBUG oslo_vmware.api [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294416, 'name': ReconfigVM_Task, 'duration_secs': 0.2887} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.607073] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Reconfigured VM instance instance-0000006f to attach disk [datastore1] 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4/5ee0e2b0-289d-47aa-b1a9-0db809ebbab4.vmdk or device None with type sparse {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1239.607637] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-025b038f-3a62-4332-ad34-29857c639f70 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.613135] env[63021]: DEBUG oslo_vmware.api [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1239.613135] env[63021]: value = "task-1294417" [ 1239.613135] env[63021]: _type = "Task" [ 1239.613135] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.620155] env[63021]: DEBUG oslo_vmware.api [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294417, 'name': Rename_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.122568] env[63021]: DEBUG oslo_vmware.api [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294417, 'name': Rename_Task, 'duration_secs': 0.136571} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.122859] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Powering on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1240.123225] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a1150734-ab87-46ea-bf4e-347601f5fff1 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.128799] env[63021]: DEBUG oslo_vmware.api [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1240.128799] env[63021]: value = "task-1294418" [ 1240.128799] env[63021]: _type = "Task" [ 1240.128799] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.135741] env[63021]: DEBUG oslo_vmware.api [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294418, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.638088] env[63021]: DEBUG oslo_vmware.api [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294418, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.138929] env[63021]: DEBUG oslo_vmware.api [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294418, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.639743] env[63021]: DEBUG oslo_vmware.api [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294418, 'name': PowerOnVM_Task, 'duration_secs': 1.157657} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1241.640104] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Powered on the VM {{(pid=63021) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1241.640230] env[63021]: INFO nova.compute.manager [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Took 9.13 seconds to spawn the instance on the hypervisor. [ 1241.640411] env[63021]: DEBUG nova.compute.manager [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Checking state {{(pid=63021) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1241.641176] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f997b27-9cad-4d31-a37d-3779571137fb {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.158475] env[63021]: INFO nova.compute.manager [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Took 14.52 seconds to build instance. [ 1242.661084] env[63021]: DEBUG oslo_concurrency.lockutils [None req-219fc1f5-de85-45c9-b392-c17efe33839a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "5ee0e2b0-289d-47aa-b1a9-0db809ebbab4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.027s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1242.992141] env[63021]: DEBUG nova.compute.manager [req-86eca677-20eb-4005-abad-1d5ece2cca11 req-2d08dab9-005f-4b33-bf89-0753394488a9 service nova] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Received event network-changed-22b596bf-eee5-4b76-b9e8-ae3dd028c5ed {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1242.992141] env[63021]: DEBUG nova.compute.manager [req-86eca677-20eb-4005-abad-1d5ece2cca11 req-2d08dab9-005f-4b33-bf89-0753394488a9 service nova] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Refreshing instance network info cache due to event network-changed-22b596bf-eee5-4b76-b9e8-ae3dd028c5ed. {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1242.992141] env[63021]: DEBUG oslo_concurrency.lockutils [req-86eca677-20eb-4005-abad-1d5ece2cca11 req-2d08dab9-005f-4b33-bf89-0753394488a9 service nova] Acquiring lock "refresh_cache-5ee0e2b0-289d-47aa-b1a9-0db809ebbab4" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1242.992141] env[63021]: DEBUG oslo_concurrency.lockutils [req-86eca677-20eb-4005-abad-1d5ece2cca11 req-2d08dab9-005f-4b33-bf89-0753394488a9 service nova] Acquired lock "refresh_cache-5ee0e2b0-289d-47aa-b1a9-0db809ebbab4" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1242.992141] env[63021]: DEBUG nova.network.neutron [req-86eca677-20eb-4005-abad-1d5ece2cca11 req-2d08dab9-005f-4b33-bf89-0753394488a9 service nova] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Refreshing network info cache for port 22b596bf-eee5-4b76-b9e8-ae3dd028c5ed {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1243.700931] env[63021]: DEBUG nova.network.neutron [req-86eca677-20eb-4005-abad-1d5ece2cca11 req-2d08dab9-005f-4b33-bf89-0753394488a9 service nova] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Updated VIF entry in instance network info cache for port 22b596bf-eee5-4b76-b9e8-ae3dd028c5ed. {{(pid=63021) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1243.701331] env[63021]: DEBUG nova.network.neutron [req-86eca677-20eb-4005-abad-1d5ece2cca11 req-2d08dab9-005f-4b33-bf89-0753394488a9 service nova] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Updating instance_info_cache with network_info: [{"id": "22b596bf-eee5-4b76-b9e8-ae3dd028c5ed", "address": "fa:16:3e:a5:25:a9", "network": {"id": "d27bb3cc-fb80-4b03-9288-ed17a0d05140", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-122910456-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.178", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "547d9a3099b94ff488ef6929ae591bac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1c797172-a569-458e-aeb0-3f21e589a740", "external-id": "nsx-vlan-transportzone-957", "segmentation_id": 957, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22b596bf-ee", "ovs_interfaceid": "22b596bf-eee5-4b76-b9e8-ae3dd028c5ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1244.203810] env[63021]: DEBUG oslo_concurrency.lockutils [req-86eca677-20eb-4005-abad-1d5ece2cca11 req-2d08dab9-005f-4b33-bf89-0753394488a9 service nova] Releasing lock "refresh_cache-5ee0e2b0-289d-47aa-b1a9-0db809ebbab4" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1272.599990] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1272.600442] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63021) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1275.595374] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1275.598995] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1275.599160] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Starting heal instance info cache {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1275.599277] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Rebuilding the list of instances to heal {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1276.128878] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquiring lock "refresh_cache-5ee0e2b0-289d-47aa-b1a9-0db809ebbab4" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1276.129041] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquired lock "refresh_cache-5ee0e2b0-289d-47aa-b1a9-0db809ebbab4" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1276.129248] env[63021]: DEBUG nova.network.neutron [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Forcefully refreshing network info cache for instance {{(pid=63021) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1276.129415] env[63021]: DEBUG nova.objects.instance [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lazy-loading 'info_cache' on Instance uuid 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1278.005342] env[63021]: DEBUG nova.network.neutron [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Updating instance_info_cache with network_info: [{"id": "22b596bf-eee5-4b76-b9e8-ae3dd028c5ed", "address": "fa:16:3e:a5:25:a9", "network": {"id": "d27bb3cc-fb80-4b03-9288-ed17a0d05140", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-122910456-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.178", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "547d9a3099b94ff488ef6929ae591bac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1c797172-a569-458e-aeb0-3f21e589a740", "external-id": "nsx-vlan-transportzone-957", "segmentation_id": 957, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22b596bf-ee", "ovs_interfaceid": "22b596bf-eee5-4b76-b9e8-ae3dd028c5ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1278.507832] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Releasing lock "refresh_cache-5ee0e2b0-289d-47aa-b1a9-0db809ebbab4" {{(pid=63021) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1278.508093] env[63021]: DEBUG nova.compute.manager [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Updated the network info_cache for instance {{(pid=63021) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1278.508315] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1278.508481] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1278.508633] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1279.012214] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1279.012591] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1279.012637] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1279.012795] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63021) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1279.013717] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72edff4f-a31d-4bb2-b6fc-8d25f0fc6826 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.022242] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa588b29-b25f-48b2-a94e-c2dd166dc8e9 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.036795] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d861d1a6-3445-4495-b79c-c64ba238de9c {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.042948] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-505bbe76-07e5-4df4-b8af-5565dca3e85d {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.071848] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181142MB free_disk=139GB free_vcpus=48 pci_devices=None {{(pid=63021) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1279.072030] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1279.072192] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1280.097087] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Instance 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63021) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1280.097312] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=63021) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1280.097439] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=63021) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1280.122075] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d409e680-cd62-4f57-9141-6746e7f51d75 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.129657] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61264f68-96d3-4018-9086-eb2210334961 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.158257] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0dcf74a-41be-4c0c-8725-369d41bc5e31 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.164840] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-824b013a-2e61-49a5-956b-4021aa7b9d30 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.177424] env[63021]: DEBUG nova.compute.provider_tree [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1280.398706] env[63021]: DEBUG oslo_concurrency.lockutils [None req-609a81d2-d361-4de4-954b-67ce7c0d583a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "5ee0e2b0-289d-47aa-b1a9-0db809ebbab4" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1280.400087] env[63021]: DEBUG oslo_concurrency.lockutils [None req-609a81d2-d361-4de4-954b-67ce7c0d583a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "5ee0e2b0-289d-47aa-b1a9-0db809ebbab4" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1280.680775] env[63021]: DEBUG nova.scheduler.client.report [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1280.902540] env[63021]: DEBUG nova.compute.utils [None req-609a81d2-d361-4de4-954b-67ce7c0d583a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Using /dev/sd instead of None {{(pid=63021) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1281.185122] env[63021]: DEBUG nova.compute.resource_tracker [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63021) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1281.185504] env[63021]: DEBUG oslo_concurrency.lockutils [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.113s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1281.276383] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1281.276595] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1281.405802] env[63021]: DEBUG oslo_concurrency.lockutils [None req-609a81d2-d361-4de4-954b-67ce7c0d583a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "5ee0e2b0-289d-47aa-b1a9-0db809ebbab4" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1281.599849] env[63021]: DEBUG oslo_service.periodic_task [None req-ed9181c4-976b-464b-be79-097e8568b5f4 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63021) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1282.471995] env[63021]: DEBUG oslo_concurrency.lockutils [None req-609a81d2-d361-4de4-954b-67ce7c0d583a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "5ee0e2b0-289d-47aa-b1a9-0db809ebbab4" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1282.472385] env[63021]: DEBUG oslo_concurrency.lockutils [None req-609a81d2-d361-4de4-954b-67ce7c0d583a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "5ee0e2b0-289d-47aa-b1a9-0db809ebbab4" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1282.472533] env[63021]: INFO nova.compute.manager [None req-609a81d2-d361-4de4-954b-67ce7c0d583a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Attaching volume 55eb526f-5104-4b04-b333-844eba2c16c7 to /dev/sdb [ 1282.501905] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a377dd4c-2977-42c2-b963-88b933316ec0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.509138] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f428de7-855a-42f2-9718-6975fbd07a2a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.521748] env[63021]: DEBUG nova.virt.block_device [None req-609a81d2-d361-4de4-954b-67ce7c0d583a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Updating existing volume attachment record: a4567f7d-765a-4999-b80a-4f215f6090ca {{(pid=63021) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1287.063351] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-609a81d2-d361-4de4-954b-67ce7c0d583a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Volume attach. Driver type: vmdk {{(pid=63021) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1287.063616] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-609a81d2-d361-4de4-954b-67ce7c0d583a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277608', 'volume_id': '55eb526f-5104-4b04-b333-844eba2c16c7', 'name': 'volume-55eb526f-5104-4b04-b333-844eba2c16c7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5ee0e2b0-289d-47aa-b1a9-0db809ebbab4', 'attached_at': '', 'detached_at': '', 'volume_id': '55eb526f-5104-4b04-b333-844eba2c16c7', 'serial': '55eb526f-5104-4b04-b333-844eba2c16c7'} {{(pid=63021) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1287.064497] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b212118b-9323-4cbd-8957-02af3ca9a45b {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.080542] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68cacd57-4cad-43ae-aa96-945ce86ad1ac {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.103711] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-609a81d2-d361-4de4-954b-67ce7c0d583a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] volume-55eb526f-5104-4b04-b333-844eba2c16c7/volume-55eb526f-5104-4b04-b333-844eba2c16c7.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1287.103954] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8045adae-7a81-4dfe-af11-f6ee0b58ffe8 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.120360] env[63021]: DEBUG oslo_vmware.api [None req-609a81d2-d361-4de4-954b-67ce7c0d583a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1287.120360] env[63021]: value = "task-1294421" [ 1287.120360] env[63021]: _type = "Task" [ 1287.120360] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1287.127478] env[63021]: DEBUG oslo_vmware.api [None req-609a81d2-d361-4de4-954b-67ce7c0d583a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294421, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1287.631824] env[63021]: DEBUG oslo_vmware.api [None req-609a81d2-d361-4de4-954b-67ce7c0d583a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294421, 'name': ReconfigVM_Task, 'duration_secs': 0.33721} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1287.632121] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-609a81d2-d361-4de4-954b-67ce7c0d583a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Reconfigured VM instance instance-0000006f to attach disk [datastore1] volume-55eb526f-5104-4b04-b333-844eba2c16c7/volume-55eb526f-5104-4b04-b333-844eba2c16c7.vmdk or device None with type thin {{(pid=63021) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1287.636650] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-063770c1-b7d4-40ec-8b90-b94e77f66a16 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.651172] env[63021]: DEBUG oslo_vmware.api [None req-609a81d2-d361-4de4-954b-67ce7c0d583a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1287.651172] env[63021]: value = "task-1294422" [ 1287.651172] env[63021]: _type = "Task" [ 1287.651172] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1287.658379] env[63021]: DEBUG oslo_vmware.api [None req-609a81d2-d361-4de4-954b-67ce7c0d583a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294422, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1288.161122] env[63021]: DEBUG oslo_vmware.api [None req-609a81d2-d361-4de4-954b-67ce7c0d583a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294422, 'name': ReconfigVM_Task, 'duration_secs': 0.132848} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1288.161496] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-609a81d2-d361-4de4-954b-67ce7c0d583a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277608', 'volume_id': '55eb526f-5104-4b04-b333-844eba2c16c7', 'name': 'volume-55eb526f-5104-4b04-b333-844eba2c16c7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5ee0e2b0-289d-47aa-b1a9-0db809ebbab4', 'attached_at': '', 'detached_at': '', 'volume_id': '55eb526f-5104-4b04-b333-844eba2c16c7', 'serial': '55eb526f-5104-4b04-b333-844eba2c16c7'} {{(pid=63021) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1289.195810] env[63021]: DEBUG nova.objects.instance [None req-609a81d2-d361-4de4-954b-67ce7c0d583a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lazy-loading 'flavor' on Instance uuid 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1289.702419] env[63021]: DEBUG oslo_concurrency.lockutils [None req-609a81d2-d361-4de4-954b-67ce7c0d583a tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "5ee0e2b0-289d-47aa-b1a9-0db809ebbab4" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.230s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1289.875128] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b0d67e58-2666-4e03-87ec-c223a647b61b tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "5ee0e2b0-289d-47aa-b1a9-0db809ebbab4" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1289.875416] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b0d67e58-2666-4e03-87ec-c223a647b61b tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "5ee0e2b0-289d-47aa-b1a9-0db809ebbab4" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1290.379044] env[63021]: INFO nova.compute.manager [None req-b0d67e58-2666-4e03-87ec-c223a647b61b tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Detaching volume 55eb526f-5104-4b04-b333-844eba2c16c7 [ 1290.408124] env[63021]: INFO nova.virt.block_device [None req-b0d67e58-2666-4e03-87ec-c223a647b61b tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Attempting to driver detach volume 55eb526f-5104-4b04-b333-844eba2c16c7 from mountpoint /dev/sdb [ 1290.408382] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0d67e58-2666-4e03-87ec-c223a647b61b tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Volume detach. Driver type: vmdk {{(pid=63021) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1290.408566] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0d67e58-2666-4e03-87ec-c223a647b61b tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277608', 'volume_id': '55eb526f-5104-4b04-b333-844eba2c16c7', 'name': 'volume-55eb526f-5104-4b04-b333-844eba2c16c7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5ee0e2b0-289d-47aa-b1a9-0db809ebbab4', 'attached_at': '', 'detached_at': '', 'volume_id': '55eb526f-5104-4b04-b333-844eba2c16c7', 'serial': '55eb526f-5104-4b04-b333-844eba2c16c7'} {{(pid=63021) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1290.409490] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2caf9ae4-bb88-4f29-962a-4324f6b261c4 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.430726] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9936c733-14f8-4f33-9027-16c6fda58618 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.437044] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88b6e41e-89ad-46a7-ac90-0adc30e918a5 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.457239] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36d5699c-58ad-46d8-9c5d-1ad53c41d719 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.471012] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0d67e58-2666-4e03-87ec-c223a647b61b tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] The volume has not been displaced from its original location: [datastore1] volume-55eb526f-5104-4b04-b333-844eba2c16c7/volume-55eb526f-5104-4b04-b333-844eba2c16c7.vmdk. No consolidation needed. {{(pid=63021) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1290.476073] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0d67e58-2666-4e03-87ec-c223a647b61b tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Reconfiguring VM instance instance-0000006f to detach disk 2001 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1290.476333] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7e5ffb16-fecc-4d8b-a21a-006c7555da27 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.493076] env[63021]: DEBUG oslo_vmware.api [None req-b0d67e58-2666-4e03-87ec-c223a647b61b tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1290.493076] env[63021]: value = "task-1294423" [ 1290.493076] env[63021]: _type = "Task" [ 1290.493076] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1290.500104] env[63021]: DEBUG oslo_vmware.api [None req-b0d67e58-2666-4e03-87ec-c223a647b61b tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294423, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1291.002587] env[63021]: DEBUG oslo_vmware.api [None req-b0d67e58-2666-4e03-87ec-c223a647b61b tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294423, 'name': ReconfigVM_Task, 'duration_secs': 0.205174} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1291.002875] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0d67e58-2666-4e03-87ec-c223a647b61b tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Reconfigured VM instance instance-0000006f to detach disk 2001 {{(pid=63021) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1291.007480] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-35f2b7c0-7332-4c34-925e-8b0b1010cf2f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.021994] env[63021]: DEBUG oslo_vmware.api [None req-b0d67e58-2666-4e03-87ec-c223a647b61b tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1291.021994] env[63021]: value = "task-1294424" [ 1291.021994] env[63021]: _type = "Task" [ 1291.021994] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1291.029176] env[63021]: DEBUG oslo_vmware.api [None req-b0d67e58-2666-4e03-87ec-c223a647b61b tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294424, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1291.531558] env[63021]: DEBUG oslo_vmware.api [None req-b0d67e58-2666-4e03-87ec-c223a647b61b tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294424, 'name': ReconfigVM_Task, 'duration_secs': 0.124759} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1291.531910] env[63021]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0d67e58-2666-4e03-87ec-c223a647b61b tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277608', 'volume_id': '55eb526f-5104-4b04-b333-844eba2c16c7', 'name': 'volume-55eb526f-5104-4b04-b333-844eba2c16c7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5ee0e2b0-289d-47aa-b1a9-0db809ebbab4', 'attached_at': '', 'detached_at': '', 'volume_id': '55eb526f-5104-4b04-b333-844eba2c16c7', 'serial': '55eb526f-5104-4b04-b333-844eba2c16c7'} {{(pid=63021) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1292.073064] env[63021]: DEBUG nova.objects.instance [None req-b0d67e58-2666-4e03-87ec-c223a647b61b tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lazy-loading 'flavor' on Instance uuid 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1293.080561] env[63021]: DEBUG oslo_concurrency.lockutils [None req-b0d67e58-2666-4e03-87ec-c223a647b61b tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "5ee0e2b0-289d-47aa-b1a9-0db809ebbab4" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.205s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1294.103257] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "5ee0e2b0-289d-47aa-b1a9-0db809ebbab4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1294.103554] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "5ee0e2b0-289d-47aa-b1a9-0db809ebbab4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1294.103766] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "5ee0e2b0-289d-47aa-b1a9-0db809ebbab4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1294.103953] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "5ee0e2b0-289d-47aa-b1a9-0db809ebbab4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1294.104149] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "5ee0e2b0-289d-47aa-b1a9-0db809ebbab4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1294.106381] env[63021]: INFO nova.compute.manager [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Terminating instance [ 1294.108205] env[63021]: DEBUG nova.compute.manager [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Start destroying the instance on the hypervisor. {{(pid=63021) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1294.108398] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Destroying instance {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1294.109249] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f717be0-a9f2-478b-903d-e9daa030fe86 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.117404] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Powering off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1294.117626] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-88f6ecb2-3c6a-4340-957e-fe034f85ef91 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.122895] env[63021]: DEBUG oslo_vmware.api [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1294.122895] env[63021]: value = "task-1294425" [ 1294.122895] env[63021]: _type = "Task" [ 1294.122895] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1294.130621] env[63021]: DEBUG oslo_vmware.api [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294425, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1294.633259] env[63021]: DEBUG oslo_vmware.api [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294425, 'name': PowerOffVM_Task, 'duration_secs': 0.218055} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1294.633529] env[63021]: DEBUG nova.virt.vmwareapi.vm_util [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Powered off the VM {{(pid=63021) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1294.633687] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Unregistering the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1294.633930] env[63021]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5afb85d1-6774-4505-a5b3-96e1613fb3ae {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.820668] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Unregistered the VM {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1294.820912] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Deleting contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1294.821124] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Deleting the datastore file [datastore1] 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4 {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1294.821429] env[63021]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-48c6d91d-0d5f-4ba5-bac8-9243bbbabaaa {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.829157] env[63021]: DEBUG oslo_vmware.api [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for the task: (returnval){ [ 1294.829157] env[63021]: value = "task-1294427" [ 1294.829157] env[63021]: _type = "Task" [ 1294.829157] env[63021]: } to complete. {{(pid=63021) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1294.836163] env[63021]: DEBUG oslo_vmware.api [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294427, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1295.340303] env[63021]: DEBUG oslo_vmware.api [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Task: {'id': task-1294427, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146208} completed successfully. {{(pid=63021) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1295.340691] env[63021]: DEBUG nova.virt.vmwareapi.ds_util [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Deleted the datastore file {{(pid=63021) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1295.340739] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Deleted contents of the VM from datastore datastore1 {{(pid=63021) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1295.340893] env[63021]: DEBUG nova.virt.vmwareapi.vmops [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Instance destroyed {{(pid=63021) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1295.341074] env[63021]: INFO nova.compute.manager [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Took 1.23 seconds to destroy the instance on the hypervisor. [ 1295.341356] env[63021]: DEBUG oslo.service.loopingcall [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63021) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1295.341589] env[63021]: DEBUG nova.compute.manager [-] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Deallocating network for instance {{(pid=63021) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1295.341686] env[63021]: DEBUG nova.network.neutron [-] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] deallocate_for_instance() {{(pid=63021) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1295.778084] env[63021]: DEBUG nova.compute.manager [req-94c40531-a5ef-485c-9461-27e328d3b7fe req-c135f1ca-eb5e-4024-a971-c9ec52157883 service nova] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Received event network-vif-deleted-22b596bf-eee5-4b76-b9e8-ae3dd028c5ed {{(pid=63021) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1295.778277] env[63021]: INFO nova.compute.manager [req-94c40531-a5ef-485c-9461-27e328d3b7fe req-c135f1ca-eb5e-4024-a971-c9ec52157883 service nova] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Neutron deleted interface 22b596bf-eee5-4b76-b9e8-ae3dd028c5ed; detaching it from the instance and deleting it from the info cache [ 1295.778457] env[63021]: DEBUG nova.network.neutron [req-94c40531-a5ef-485c-9461-27e328d3b7fe req-c135f1ca-eb5e-4024-a971-c9ec52157883 service nova] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1296.253455] env[63021]: DEBUG nova.network.neutron [-] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Updating instance_info_cache with network_info: [] {{(pid=63021) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1296.281552] env[63021]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eb7b8f3c-7f2d-4e71-95bc-ab8c4c6dc4f0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.291376] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77828f2a-0749-4b6b-a4f7-6357536e0f01 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.313811] env[63021]: DEBUG nova.compute.manager [req-94c40531-a5ef-485c-9461-27e328d3b7fe req-c135f1ca-eb5e-4024-a971-c9ec52157883 service nova] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Detach interface failed, port_id=22b596bf-eee5-4b76-b9e8-ae3dd028c5ed, reason: Instance 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4 could not be found. {{(pid=63021) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1296.756642] env[63021]: INFO nova.compute.manager [-] [instance: 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4] Took 1.41 seconds to deallocate network for instance. [ 1297.263025] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1297.263346] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1297.263575] env[63021]: DEBUG nova.objects.instance [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lazy-loading 'resources' on Instance uuid 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4 {{(pid=63021) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1297.796956] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edb9b925-8444-4a9f-9bfc-02d940cf0475 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.804464] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e68ffab3-7b36-4db6-bc93-22288e3f747a {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.834110] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0f64321-03a9-4d20-ad15-2ad8ce18ff6f {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.840901] env[63021]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfc139a7-753a-4a03-9e47-1bb2c82ef6c0 {{(pid=63021) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.853408] env[63021]: DEBUG nova.compute.provider_tree [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Inventory has not changed in ProviderTree for provider: 6ef6c041-2b0d-4bc3-8662-5d679a77a86d {{(pid=63021) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1298.356799] env[63021]: DEBUG nova.scheduler.client.report [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Inventory has not changed for provider 6ef6c041-2b0d-4bc3-8662-5d679a77a86d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63021) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1298.862510] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.599s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1298.884824] env[63021]: INFO nova.scheduler.client.report [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Deleted allocations for instance 5ee0e2b0-289d-47aa-b1a9-0db809ebbab4 [ 1299.395179] env[63021]: DEBUG oslo_concurrency.lockutils [None req-c892e0e8-2879-4175-8e61-4e5ba3c81092 tempest-AttachVolumeNegativeTest-2037961347 tempest-AttachVolumeNegativeTest-2037961347-project-member] Lock "5ee0e2b0-289d-47aa-b1a9-0db809ebbab4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.291s {{(pid=63021) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}}